var/home/core/zuul-output/0000755000175000017500000000000015114536100014521 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114563210015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006721547315114563201017712 0ustar rootrootDec 05 11:08:23 crc systemd[1]: Starting Kubernetes Kubelet... Dec 05 11:08:23 crc restorecon[4808]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:23 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 11:08:24 crc restorecon[4808]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 05 11:08:24 crc kubenswrapper[4809]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 11:08:24 crc kubenswrapper[4809]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 11:08:24 crc kubenswrapper[4809]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 11:08:24 crc kubenswrapper[4809]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 11:08:24 crc kubenswrapper[4809]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 11:08:24 crc kubenswrapper[4809]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.726552 4809 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730445 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730469 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730475 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730480 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730484 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730488 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730494 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730499 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730503 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730508 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730514 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730520 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730526 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730531 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730536 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730540 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730546 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730558 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730563 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730567 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730571 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730577 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730582 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730586 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730590 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730594 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730598 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730603 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730608 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730615 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730620 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730645 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730649 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730654 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730658 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730662 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730666 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730670 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730674 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730679 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730683 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730688 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730695 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730700 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730704 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730708 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730712 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730716 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730721 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730725 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730729 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730733 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730737 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730741 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730747 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730751 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730755 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730759 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730763 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730767 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730771 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730775 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730778 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730782 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730785 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730789 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730793 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730797 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730801 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730805 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.730810 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730926 4809 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730936 4809 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730946 4809 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730954 4809 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730961 4809 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730966 4809 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730973 4809 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730981 4809 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730987 4809 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730992 4809 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.730998 4809 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731004 4809 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731010 4809 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731014 4809 flags.go:64] FLAG: --cgroup-root="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731018 4809 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731022 4809 flags.go:64] FLAG: --client-ca-file="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731026 4809 flags.go:64] FLAG: --cloud-config="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731030 4809 flags.go:64] FLAG: --cloud-provider="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731034 4809 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731039 4809 flags.go:64] FLAG: --cluster-domain="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731044 4809 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731049 4809 flags.go:64] FLAG: --config-dir="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731052 4809 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731057 4809 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731063 4809 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731067 4809 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731072 4809 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731076 4809 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731080 4809 flags.go:64] FLAG: --contention-profiling="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731084 4809 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731089 4809 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731093 4809 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731097 4809 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731102 4809 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731106 4809 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731111 4809 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731115 4809 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731119 4809 flags.go:64] FLAG: --enable-server="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731123 4809 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731130 4809 flags.go:64] FLAG: --event-burst="100" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731135 4809 flags.go:64] FLAG: --event-qps="50" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731139 4809 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731143 4809 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731149 4809 flags.go:64] FLAG: --eviction-hard="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731154 4809 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731158 4809 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731162 4809 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731167 4809 flags.go:64] FLAG: --eviction-soft="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731171 4809 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731175 4809 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731179 4809 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731183 4809 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731193 4809 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731197 4809 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731201 4809 flags.go:64] FLAG: --feature-gates="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731206 4809 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731211 4809 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731215 4809 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731219 4809 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731224 4809 flags.go:64] FLAG: --healthz-port="10248" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731228 4809 flags.go:64] FLAG: --help="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731234 4809 flags.go:64] FLAG: --hostname-override="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731239 4809 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731244 4809 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731250 4809 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731254 4809 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731260 4809 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731265 4809 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731272 4809 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731277 4809 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731282 4809 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731287 4809 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731293 4809 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731299 4809 flags.go:64] FLAG: --kube-reserved="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731306 4809 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731312 4809 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731317 4809 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731322 4809 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731328 4809 flags.go:64] FLAG: --lock-file="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731334 4809 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731340 4809 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731345 4809 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731362 4809 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731367 4809 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731375 4809 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731380 4809 flags.go:64] FLAG: --logging-format="text" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731385 4809 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731391 4809 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731396 4809 flags.go:64] FLAG: --manifest-url="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731401 4809 flags.go:64] FLAG: --manifest-url-header="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731408 4809 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731412 4809 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731418 4809 flags.go:64] FLAG: --max-pods="110" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731422 4809 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731427 4809 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731431 4809 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731435 4809 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731440 4809 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731444 4809 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731449 4809 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731459 4809 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731463 4809 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731467 4809 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731472 4809 flags.go:64] FLAG: --pod-cidr="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731476 4809 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731482 4809 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731486 4809 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731491 4809 flags.go:64] FLAG: --pods-per-core="0" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731497 4809 flags.go:64] FLAG: --port="10250" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731502 4809 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731507 4809 flags.go:64] FLAG: --provider-id="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731513 4809 flags.go:64] FLAG: --qos-reserved="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731518 4809 flags.go:64] FLAG: --read-only-port="10255" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731524 4809 flags.go:64] FLAG: --register-node="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731529 4809 flags.go:64] FLAG: --register-schedulable="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731535 4809 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731542 4809 flags.go:64] FLAG: --registry-burst="10" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731548 4809 flags.go:64] FLAG: --registry-qps="5" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731553 4809 flags.go:64] FLAG: --reserved-cpus="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731557 4809 flags.go:64] FLAG: --reserved-memory="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731563 4809 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731567 4809 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731572 4809 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731576 4809 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731580 4809 flags.go:64] FLAG: --runonce="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731584 4809 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731588 4809 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731593 4809 flags.go:64] FLAG: --seccomp-default="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731597 4809 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731601 4809 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731605 4809 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731610 4809 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731615 4809 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731619 4809 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731639 4809 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731644 4809 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731648 4809 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731653 4809 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731657 4809 flags.go:64] FLAG: --system-cgroups="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731662 4809 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731669 4809 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731673 4809 flags.go:64] FLAG: --tls-cert-file="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731677 4809 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731690 4809 flags.go:64] FLAG: --tls-min-version="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731695 4809 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731700 4809 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731704 4809 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731708 4809 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731714 4809 flags.go:64] FLAG: --v="2" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731721 4809 flags.go:64] FLAG: --version="false" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731727 4809 flags.go:64] FLAG: --vmodule="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731733 4809 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.731738 4809 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731856 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731860 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731865 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731868 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731872 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731876 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731880 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731884 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731888 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731891 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731895 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731899 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731904 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731908 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731912 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731916 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731920 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731924 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731941 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731947 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731952 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731956 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731961 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731966 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731971 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731975 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731980 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731984 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731989 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731992 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731996 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.731999 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732003 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732006 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732012 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732016 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732020 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732024 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732027 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732031 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732035 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732041 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732045 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732048 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732052 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732056 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732060 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732063 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732067 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732070 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732074 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732078 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732082 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732087 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732092 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732096 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732106 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732110 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732115 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732119 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732128 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732133 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732137 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732141 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732145 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732149 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732152 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732156 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732159 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732163 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.732167 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.732178 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.742167 4809 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.742207 4809 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742295 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742309 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742313 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742318 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742322 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742325 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742329 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742332 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742336 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742340 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742343 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742347 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742350 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742354 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742359 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742363 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742368 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742372 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742376 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742381 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742386 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742390 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742394 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742398 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742403 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742407 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742411 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742417 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742423 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742428 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742432 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742438 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742442 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742446 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742452 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742456 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742461 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742467 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742473 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742478 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742482 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742487 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742491 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742494 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742498 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742503 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742507 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742510 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742514 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742517 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742521 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742525 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742530 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742534 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742539 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742543 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742547 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742550 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742554 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742558 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742562 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742566 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742569 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742573 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742576 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742580 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742584 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742587 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742591 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742594 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742598 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.742605 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742745 4809 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742754 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742758 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742763 4809 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742767 4809 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742771 4809 feature_gate.go:330] unrecognized feature gate: Example Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742775 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742779 4809 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742783 4809 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742786 4809 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742791 4809 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742795 4809 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742799 4809 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742804 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742809 4809 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742813 4809 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742817 4809 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742821 4809 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742826 4809 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742830 4809 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742834 4809 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742873 4809 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742878 4809 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742882 4809 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742885 4809 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742889 4809 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742894 4809 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742899 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742902 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742906 4809 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742911 4809 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742915 4809 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742918 4809 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742922 4809 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742926 4809 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742930 4809 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742934 4809 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742937 4809 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742941 4809 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742944 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742948 4809 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742952 4809 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742955 4809 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742958 4809 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742962 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742965 4809 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742969 4809 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742973 4809 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742976 4809 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742979 4809 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742983 4809 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742986 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742990 4809 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742993 4809 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.742998 4809 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743002 4809 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743006 4809 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743009 4809 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743014 4809 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743018 4809 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743022 4809 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743025 4809 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743029 4809 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743032 4809 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743036 4809 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743039 4809 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743043 4809 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743046 4809 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743050 4809 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743053 4809 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.743057 4809 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.743063 4809 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.743436 4809 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.745670 4809 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.745749 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.746137 4809 server.go:997] "Starting client certificate rotation" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.746155 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.746455 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-05 05:42:26.784809273 +0000 UTC Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.746792 4809 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 738h34m2.03802563s for next certificate rotation Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.750569 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.751959 4809 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.761034 4809 log.go:25] "Validated CRI v1 runtime API" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.772781 4809 log.go:25] "Validated CRI v1 image API" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.774578 4809 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.777437 4809 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-05-11-03-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.777477 4809 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.798374 4809 manager.go:217] Machine: {Timestamp:2025-12-05 11:08:24.796957282 +0000 UTC m=+0.187933860 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:67643894-7483-44e6-a6f9-98b40195f076 BootID:2141eb26-0a7e-4c21-87d5-ee9ba8842777 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:bf:01:bf Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:bf:01:bf Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ae:9c:13 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:23:a4:be Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a7:fc:98 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:6a:f8:d0 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:a2:e1:a5 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ee:56:7a:ee:35:93 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:92:fa:37:e8:88:5b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.798656 4809 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.798851 4809 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.799510 4809 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.799735 4809 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.799777 4809 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.800006 4809 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.800018 4809 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.800221 4809 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.800251 4809 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.800521 4809 state_mem.go:36] "Initialized new in-memory state store" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.800606 4809 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.801385 4809 kubelet.go:418] "Attempting to sync node with API server" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.801407 4809 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.801431 4809 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.801445 4809 kubelet.go:324] "Adding apiserver pod source" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.801458 4809 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.803338 4809 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.804180 4809 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805093 4809 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.805074 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.805208 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.805293 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.805380 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805864 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805899 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805910 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805921 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805938 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805948 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805958 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805976 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.805989 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.806000 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.806016 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.806026 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.806071 4809 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.806746 4809 server.go:1280] "Started kubelet" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.808414 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.809591 4809 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.809715 4809 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.810337 4809 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.810251 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e4d203b669f2c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 11:08:24.806686508 +0000 UTC m=+0.197663076,LastTimestamp:2025-12-05 11:08:24.806686508 +0000 UTC m=+0.197663076,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 11:08:24 crc systemd[1]: Started Kubernetes Kubelet. Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.813308 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.813387 4809 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.815572 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 18:03:28.778786959 +0000 UTC Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.815688 4809 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.815674 4809 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.815722 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.815772 4809 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.816217 4809 server.go:460] "Adding debug handlers to kubelet server" Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.816278 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.818076 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.819213 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.819447 4809 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.819465 4809 factory.go:55] Registering systemd factory Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.819473 4809 factory.go:221] Registration of the systemd container factory successfully Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.819788 4809 factory.go:153] Registering CRI-O factory Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.819815 4809 factory.go:221] Registration of the crio container factory successfully Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.819852 4809 factory.go:103] Registering Raw factory Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.819873 4809 manager.go:1196] Started watching for new ooms in manager Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.820477 4809 manager.go:319] Starting recovery of all containers Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.834995 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.835320 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.835425 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.835525 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.835717 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.835860 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.835983 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836117 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836226 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836323 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836421 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836506 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836594 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836730 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836827 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.836929 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837054 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837163 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837288 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837444 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837594 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837742 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837858 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.837993 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.838103 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.838213 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.838355 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.838457 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.838561 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.838715 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.838839 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.839004 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.839112 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.839216 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840391 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840468 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840512 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840534 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840555 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840576 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840600 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840619 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840662 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840681 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840704 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840725 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840746 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840766 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840784 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840801 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840822 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840843 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840918 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840940 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840959 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840973 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.840989 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841006 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841028 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841051 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841069 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841089 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841111 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841131 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841151 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841168 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841185 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841204 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841223 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841242 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841262 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841278 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841291 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841306 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841321 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841336 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841350 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841364 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841381 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841396 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841411 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841425 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841441 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841459 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841479 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841499 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841516 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841530 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841546 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841561 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841596 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841646 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841667 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841686 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841706 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841721 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841736 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841751 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841765 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841779 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841793 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841809 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841831 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841845 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841870 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841888 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841903 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841916 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841932 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.841948 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842161 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842181 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842200 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842220 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842240 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842257 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842275 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842291 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842309 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842335 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842352 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842367 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842382 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842396 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842410 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842423 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842435 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842451 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842469 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842501 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842517 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842530 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842543 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842559 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842579 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842598 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842616 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842685 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842703 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842721 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842740 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842758 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842774 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842811 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.842828 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845086 4809 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845126 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845144 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845159 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845174 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845190 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845205 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845221 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845235 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845249 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845265 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845280 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845294 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845308 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845324 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845338 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845358 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845374 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845389 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845403 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845418 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845433 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845446 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845460 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845478 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845498 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845512 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845525 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845539 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845554 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845567 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845583 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845598 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845613 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845656 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845677 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845698 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845717 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845737 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845755 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845781 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845795 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845809 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845824 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845837 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845854 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845870 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845893 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845909 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845926 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845941 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845978 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.845993 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846009 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846023 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846037 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846051 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846066 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846081 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846095 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846110 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846126 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846141 4809 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846154 4809 reconstruct.go:97] "Volume reconstruction finished" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.846165 4809 reconciler.go:26] "Reconciler: start to sync state" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.847747 4809 manager.go:324] Recovery completed Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.858193 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.859880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.860052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.860181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.865673 4809 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.865738 4809 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.865775 4809 state_mem.go:36] "Initialized new in-memory state store" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.868566 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.870693 4809 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.870844 4809 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.870897 4809 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.871198 4809 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 05 11:08:24 crc kubenswrapper[4809]: W1205 11:08:24.873762 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.873843 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.916026 4809 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.940508 4809 policy_none.go:49] "None policy: Start" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.942078 4809 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 11:08:24 crc kubenswrapper[4809]: I1205 11:08:24.942124 4809 state_mem.go:35] "Initializing new in-memory state store" Dec 05 11:08:24 crc kubenswrapper[4809]: E1205 11:08:24.971819 4809 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005005 4809 manager.go:334] "Starting Device Plugin manager" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005230 4809 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005248 4809 server.go:79] "Starting device plugin registration server" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005675 4809 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005695 4809 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005821 4809 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005939 4809 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.005951 4809 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 11:08:25 crc kubenswrapper[4809]: E1205 11:08:25.013130 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 11:08:25 crc kubenswrapper[4809]: E1205 11:08:25.017003 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.106696 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.108509 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.108570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.108593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.108672 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 11:08:25 crc kubenswrapper[4809]: E1205 11:08:25.109330 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.173332 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.173429 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.174730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.174787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.174802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.174955 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.175662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.175769 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.176072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.176168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.176178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.176302 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.176731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.176778 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.177498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.177574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.177582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.177804 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178007 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178047 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.178854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179573 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.179897 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181755 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.181783 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.182663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.182695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.182704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249532 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249595 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249616 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249658 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249716 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249742 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249904 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249946 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.249983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.250005 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.250029 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.250050 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.250071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.310247 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.311401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.311467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.311484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.311521 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 11:08:25 crc kubenswrapper[4809]: E1205 11:08:25.312070 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351722 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351743 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351761 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351794 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351847 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351988 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351921 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351946 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351950 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351959 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351972 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.351864 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352284 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352349 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352372 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352396 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.352623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: E1205 11:08:25.418143 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.515368 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.524517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: W1205 11:08:25.545430 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f3f2c0942c42ef482718e41f3a491937becf9a1e21807ed5f9e0e315dcbfb39c WatchSource:0}: Error finding container f3f2c0942c42ef482718e41f3a491937becf9a1e21807ed5f9e0e315dcbfb39c: Status 404 returned error can't find the container with id f3f2c0942c42ef482718e41f3a491937becf9a1e21807ed5f9e0e315dcbfb39c Dec 05 11:08:25 crc kubenswrapper[4809]: W1205 11:08:25.547781 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-161940f801a17239b36c60a196c105a6ba93187546687224f6e91da619599f61 WatchSource:0}: Error finding container 161940f801a17239b36c60a196c105a6ba93187546687224f6e91da619599f61: Status 404 returned error can't find the container with id 161940f801a17239b36c60a196c105a6ba93187546687224f6e91da619599f61 Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.576585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: W1205 11:08:25.592607 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-5c92902a27861fa624ec95c21d0229bdb1f5d25ecb5f523fbfd5876a90b07e0c WatchSource:0}: Error finding container 5c92902a27861fa624ec95c21d0229bdb1f5d25ecb5f523fbfd5876a90b07e0c: Status 404 returned error can't find the container with id 5c92902a27861fa624ec95c21d0229bdb1f5d25ecb5f523fbfd5876a90b07e0c Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.605135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.611314 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:25 crc kubenswrapper[4809]: W1205 11:08:25.624899 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-08a06103b342ca5b6c1ed02bff175468e5e27fdfd26fcbc3742c0fc00e25dd19 WatchSource:0}: Error finding container 08a06103b342ca5b6c1ed02bff175468e5e27fdfd26fcbc3742c0fc00e25dd19: Status 404 returned error can't find the container with id 08a06103b342ca5b6c1ed02bff175468e5e27fdfd26fcbc3742c0fc00e25dd19 Dec 05 11:08:25 crc kubenswrapper[4809]: W1205 11:08:25.630913 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d1de15c66f2e2b2a799886f064ed4f890ba7772f2b5bd72bd94e69f6aeb49a64 WatchSource:0}: Error finding container d1de15c66f2e2b2a799886f064ed4f890ba7772f2b5bd72bd94e69f6aeb49a64: Status 404 returned error can't find the container with id d1de15c66f2e2b2a799886f064ed4f890ba7772f2b5bd72bd94e69f6aeb49a64 Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.712954 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.714345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.714392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.714405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.714436 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 11:08:25 crc kubenswrapper[4809]: E1205 11:08:25.715134 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.810141 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.816279 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 14:38:34.617372064 +0000 UTC Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.876084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f3f2c0942c42ef482718e41f3a491937becf9a1e21807ed5f9e0e315dcbfb39c"} Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.880011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d1de15c66f2e2b2a799886f064ed4f890ba7772f2b5bd72bd94e69f6aeb49a64"} Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.886515 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"08a06103b342ca5b6c1ed02bff175468e5e27fdfd26fcbc3742c0fc00e25dd19"} Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.888186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5c92902a27861fa624ec95c21d0229bdb1f5d25ecb5f523fbfd5876a90b07e0c"} Dec 05 11:08:25 crc kubenswrapper[4809]: I1205 11:08:25.890700 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"161940f801a17239b36c60a196c105a6ba93187546687224f6e91da619599f61"} Dec 05 11:08:26 crc kubenswrapper[4809]: W1205 11:08:26.008698 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:26 crc kubenswrapper[4809]: E1205 11:08:26.009141 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:26 crc kubenswrapper[4809]: W1205 11:08:26.032255 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:26 crc kubenswrapper[4809]: E1205 11:08:26.032343 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:26 crc kubenswrapper[4809]: W1205 11:08:26.034994 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:26 crc kubenswrapper[4809]: E1205 11:08:26.035038 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:26 crc kubenswrapper[4809]: E1205 11:08:26.219036 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Dec 05 11:08:26 crc kubenswrapper[4809]: W1205 11:08:26.359602 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:26 crc kubenswrapper[4809]: E1205 11:08:26.359713 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.515970 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.517459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.517520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.517532 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.517554 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 11:08:26 crc kubenswrapper[4809]: E1205 11:08:26.517978 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.809451 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.817157 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 20:14:18.745762781 +0000 UTC Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.895618 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f" exitCode=0 Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.895754 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.895864 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.896704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.896732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.896744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.902006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.902084 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.902098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.902108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.902293 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.903429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.903459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.903471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.908484 4809 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008" exitCode=0 Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.908574 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.908693 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.913378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.913453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.913468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.917797 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373" exitCode=0 Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.917883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.917985 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.919512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.919550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.919563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.920473 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8" exitCode=0 Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.920518 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8"} Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.920672 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.923248 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.923326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.923349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.927937 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.929091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.929131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:26 crc kubenswrapper[4809]: I1205 11:08:26.929146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:27 crc kubenswrapper[4809]: W1205 11:08:27.756337 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:27 crc kubenswrapper[4809]: E1205 11:08:27.756468 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:27 crc kubenswrapper[4809]: W1205 11:08:27.807059 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Dec 05 11:08:27 crc kubenswrapper[4809]: E1205 11:08:27.807177 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.817269 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 09:57:24.097958076 +0000 UTC Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.817333 4809 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 382h48m56.280627159s for next certificate rotation Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.927087 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c" exitCode=0 Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.927209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.927236 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.928650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.928781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.928904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.932139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.932173 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.932189 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.934381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0a6a8e3f6b18707b772ca7e0a55ec37e40c7ab09c75c6585b9715187f221505a"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.934489 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.935252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.935280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.935292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.938323 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.938894 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.939304 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.939335 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.939352 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6"} Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.939746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.939775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.939789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.944060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.944142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:27 crc kubenswrapper[4809]: I1205 11:08:27.944171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.118470 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.119788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.119878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.119891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.119923 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.944767 4809 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b" exitCode=0 Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.944857 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b"} Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.945022 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.946160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.946214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.946231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.951453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510"} Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.951522 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.951521 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa"} Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.951480 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.951725 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.952017 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.952706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.952758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.952782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.952872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.952902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.952913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.953793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.953906 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:28 crc kubenswrapper[4809]: I1205 11:08:28.953927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.018518 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.018733 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.020237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.020276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.020288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.956334 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0"} Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.956377 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53"} Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.956393 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3"} Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.956402 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6"} Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.956442 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.956489 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.959106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.959149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.959165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.990987 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.991134 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.992414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.992441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.992449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:29 crc kubenswrapper[4809]: I1205 11:08:29.996083 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.969582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9"} Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.969653 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.969749 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.970958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.971002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.971011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.971074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.971107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:30 crc kubenswrapper[4809]: I1205 11:08:30.971123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.160958 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.161145 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.161186 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.162237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.162269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.162280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.281258 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.566867 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.971547 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.971603 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.971675 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.971676 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:31 crc kubenswrapper[4809]: I1205 11:08:31.973322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:32 crc kubenswrapper[4809]: I1205 11:08:32.538856 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:32 crc kubenswrapper[4809]: I1205 11:08:32.974918 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:32 crc kubenswrapper[4809]: I1205 11:08:32.976064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:32 crc kubenswrapper[4809]: I1205 11:08:32.976121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:32 crc kubenswrapper[4809]: I1205 11:08:32.976139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.627577 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.627885 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.629408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.629500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.629526 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.678674 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.678944 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.680368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.680411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:33 crc kubenswrapper[4809]: I1205 11:08:33.680423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:34 crc kubenswrapper[4809]: I1205 11:08:34.187199 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:08:34 crc kubenswrapper[4809]: I1205 11:08:34.187395 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:34 crc kubenswrapper[4809]: I1205 11:08:34.188863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:34 crc kubenswrapper[4809]: I1205 11:08:34.188924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:34 crc kubenswrapper[4809]: I1205 11:08:34.188945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:35 crc kubenswrapper[4809]: E1205 11:08:35.013255 4809 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 11:08:35 crc kubenswrapper[4809]: I1205 11:08:35.539971 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 11:08:35 crc kubenswrapper[4809]: I1205 11:08:35.540060 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:08:37 crc kubenswrapper[4809]: I1205 11:08:37.810504 4809 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 05 11:08:37 crc kubenswrapper[4809]: E1205 11:08:37.820022 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 05 11:08:38 crc kubenswrapper[4809]: W1205 11:08:38.109795 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 11:08:38 crc kubenswrapper[4809]: I1205 11:08:38.109902 4809 trace.go:236] Trace[229364634]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 11:08:28.108) (total time: 10001ms): Dec 05 11:08:38 crc kubenswrapper[4809]: Trace[229364634]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (11:08:38.109) Dec 05 11:08:38 crc kubenswrapper[4809]: Trace[229364634]: [10.001519971s] [10.001519971s] END Dec 05 11:08:38 crc kubenswrapper[4809]: E1205 11:08:38.109927 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 11:08:38 crc kubenswrapper[4809]: E1205 11:08:38.121503 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 05 11:08:38 crc kubenswrapper[4809]: W1205 11:08:38.306346 4809 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 11:08:38 crc kubenswrapper[4809]: I1205 11:08:38.306513 4809 trace.go:236] Trace[1422267173]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 11:08:28.305) (total time: 10001ms): Dec 05 11:08:38 crc kubenswrapper[4809]: Trace[1422267173]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (11:08:38.306) Dec 05 11:08:38 crc kubenswrapper[4809]: Trace[1422267173]: [10.001358417s] [10.001358417s] END Dec 05 11:08:38 crc kubenswrapper[4809]: E1205 11:08:38.306554 4809 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 11:08:38 crc kubenswrapper[4809]: I1205 11:08:38.889365 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 11:08:38 crc kubenswrapper[4809]: I1205 11:08:38.889453 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 11:08:38 crc kubenswrapper[4809]: I1205 11:08:38.898730 4809 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 11:08:38 crc kubenswrapper[4809]: I1205 11:08:38.898820 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 11:08:39 crc kubenswrapper[4809]: I1205 11:08:39.025877 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:39 crc kubenswrapper[4809]: I1205 11:08:39.026045 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:39 crc kubenswrapper[4809]: I1205 11:08:39.027289 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:39 crc kubenswrapper[4809]: I1205 11:08:39.027419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:39 crc kubenswrapper[4809]: I1205 11:08:39.027584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:40 crc kubenswrapper[4809]: I1205 11:08:40.383837 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 05 11:08:40 crc kubenswrapper[4809]: I1205 11:08:40.384769 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:40 crc kubenswrapper[4809]: I1205 11:08:40.386089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:40 crc kubenswrapper[4809]: I1205 11:08:40.386198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:40 crc kubenswrapper[4809]: I1205 11:08:40.386219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:40 crc kubenswrapper[4809]: I1205 11:08:40.415950 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.000959 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.002117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.002154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.002166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.017822 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.167731 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.167901 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.169222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.169505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.169745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.175076 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.322137 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.324012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.324081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.324097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.324133 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 11:08:41 crc kubenswrapper[4809]: E1205 11:08:41.329238 4809 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.802537 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.811196 4809 apiserver.go:52] "Watching apiserver" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.814202 4809 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.814693 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.815302 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:41 crc kubenswrapper[4809]: E1205 11:08:41.815419 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.815494 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.815533 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:41 crc kubenswrapper[4809]: E1205 11:08:41.815587 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:41 crc kubenswrapper[4809]: E1205 11:08:41.815645 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.815602 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.815701 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.815665 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.816593 4809 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.819336 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.819475 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.819713 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.819721 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.819857 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.819906 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.819865 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.820210 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.820605 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.853374 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.870687 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.885584 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.900665 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.915748 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.927532 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:41 crc kubenswrapper[4809]: I1205 11:08:41.939749 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:42 crc kubenswrapper[4809]: I1205 11:08:42.020730 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 05 11:08:42 crc kubenswrapper[4809]: I1205 11:08:42.020863 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 11:08:42 crc kubenswrapper[4809]: I1205 11:08:42.325996 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.006313 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.871524 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.871524 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.871682 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:43 crc kubenswrapper[4809]: E1205 11:08:43.871759 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:43 crc kubenswrapper[4809]: E1205 11:08:43.871850 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:43 crc kubenswrapper[4809]: E1205 11:08:43.872012 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.893566 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.893939 4809 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.895606 4809 trace.go:236] Trace[397601128]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 11:08:33.066) (total time: 10828ms): Dec 05 11:08:43 crc kubenswrapper[4809]: Trace[397601128]: ---"Objects listed" error: 10828ms (11:08:43.895) Dec 05 11:08:43 crc kubenswrapper[4809]: Trace[397601128]: [10.828690818s] [10.828690818s] END Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.895626 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.970612 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.983495 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994448 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994490 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994509 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994525 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994541 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994556 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994571 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994592 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994608 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994623 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994653 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994686 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994702 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994720 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994734 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994749 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994763 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994781 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994795 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994785 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994812 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994844 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994877 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994891 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994908 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994962 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994977 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.994990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995006 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995022 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995018 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995072 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995100 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995136 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995152 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995168 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995184 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995200 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995210 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995215 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995323 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995454 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995476 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995534 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995552 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995601 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995613 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995617 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995674 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995690 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995704 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995718 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995747 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995761 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995791 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995788 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995806 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995814 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995822 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995900 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995936 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995954 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995971 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.995989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996006 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996024 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996078 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996095 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996111 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996128 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996178 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996194 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996210 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996229 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996245 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996261 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996296 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996314 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996331 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996351 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996367 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996383 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996399 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996416 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996448 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996509 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996526 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996541 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996556 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996573 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996668 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996696 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996717 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996758 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996773 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996803 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996834 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996848 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996864 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996929 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996946 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996961 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996977 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.996991 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997007 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997023 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997038 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997054 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997071 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997087 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997103 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997164 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997181 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997197 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997217 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997232 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997249 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997265 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997297 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997313 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997330 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 11:08:43 crc kubenswrapper[4809]: I1205 11:08:43.997346 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997378 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997411 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997442 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997457 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997478 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997499 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997514 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997546 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997561 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997579 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997605 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999155 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999176 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999192 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999208 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999395 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999587 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999640 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999658 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999693 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999723 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999746 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999765 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999799 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999835 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999857 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999875 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999891 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999907 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999923 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999939 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999962 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999986 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000010 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000075 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000091 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000108 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000123 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000154 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000171 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000191 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000211 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000226 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000243 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000279 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000295 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000313 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000350 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000416 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000433 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000454 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000493 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000547 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000582 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000600 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000700 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000712 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000723 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000734 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000744 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000757 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000781 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000795 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000807 4809 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.996073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.996136 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.996292 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.996333 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.996959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997370 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.010731 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997517 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997843 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.997878 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998075 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998099 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998291 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998304 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998435 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998580 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998589 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.010806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998790 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999020 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999096 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999747 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999819 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.010886 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.999948 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000138 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000295 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000321 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000441 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000734 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.000896 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.001446 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:43.998598 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.001469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.001541 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.002125 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.002330 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.002811 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.003048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.003089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.006965 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.007065 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.007265 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.007431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.007440 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.007447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.007566 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.007849 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.008068 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.008419 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.009453 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.009832 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.010143 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.001479 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.011490 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:08:44.511465389 +0000 UTC m=+19.902441947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.011545 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.011938 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.013767 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.013986 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.014056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.015119 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.015186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.015933 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.015996 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.016678 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.016967 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.017260 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.017327 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.017461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.017795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.017844 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.020235 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.020340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.020771 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.020938 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021127 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021223 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.017661 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021349 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021517 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.017101 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021786 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021937 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.021953 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022254 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022269 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022477 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022567 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022653 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022720 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022832 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022881 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.022907 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023023 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023044 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023094 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023104 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023274 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023301 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023554 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023588 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.023713 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.025179 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.025568 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.025820 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.025879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.026587 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.026722 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.026801 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.027035 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.027081 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.027200 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.027803 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.027837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.028227 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.028390 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.028727 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.029321 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.029343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.030267 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.030677 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.030822 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.030889 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.030954 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:44.53093093 +0000 UTC m=+19.921907488 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031107 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.031429 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.031474 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:44.531464845 +0000 UTC m=+19.922441403 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031542 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031676 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031780 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031903 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.031919 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.032153 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.032369 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.032765 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.032798 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.032844 4809 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.032969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.033153 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.033253 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.033287 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.033829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.034317 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.034893 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.035615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.035685 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.035990 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.036155 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.036233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.036366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.037919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.038164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.039820 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.040148 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.040488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.040493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.040745 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.042058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.045061 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.045106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.045473 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.048961 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.049075 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.056758 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.056934 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.057223 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.065859 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.065906 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.065923 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.065997 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:44.565972619 +0000 UTC m=+19.956949367 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.072978 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.073074 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.073266 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.073850 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.074346 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.074613 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.074662 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.074680 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.074733 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:44.574712613 +0000 UTC m=+19.965689351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.076786 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.076891 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.077256 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.078521 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.078577 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.078841 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.080952 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.080962 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.084309 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.084444 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.085256 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.085923 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.085962 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.086001 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.086251 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.087054 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.087070 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.091892 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.093943 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.099999 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.099991 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.100472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.101848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.101968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102137 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102219 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102295 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102366 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102427 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102502 4809 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102577 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102664 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102735 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102796 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102871 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.102947 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103053 4809 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103124 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103204 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103262 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103337 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103407 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103481 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103553 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103609 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103708 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103791 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103871 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103943 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104001 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104127 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104187 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104265 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104341 4809 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104411 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104483 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104535 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104609 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104702 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104780 4809 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104851 4809 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104910 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104982 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105035 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105093 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105170 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105226 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105279 4809 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105708 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105779 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105847 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105905 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.105965 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106023 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106092 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.104012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.103968 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106151 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106242 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106259 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106273 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106286 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106298 4809 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106309 4809 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106322 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106335 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106348 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106363 4809 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106377 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106388 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106401 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106413 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106425 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106436 4809 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106448 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106460 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106472 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106484 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106495 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106507 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106518 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106530 4809 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106544 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106558 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106572 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106586 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106599 4809 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106614 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106643 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106658 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106670 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106683 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106694 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106706 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106718 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106731 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106746 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106758 4809 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106768 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106780 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106791 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106802 4809 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106813 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106823 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106834 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106845 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106856 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106870 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106881 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106892 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106902 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106912 4809 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106923 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106933 4809 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106945 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106957 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106970 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106981 4809 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106991 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107001 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107014 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107025 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107034 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107045 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107055 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107066 4809 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107077 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107087 4809 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107099 4809 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107111 4809 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107121 4809 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107131 4809 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107142 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107153 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107164 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107175 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107187 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107198 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107210 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107221 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107231 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107243 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107255 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107268 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107280 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107293 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107305 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107316 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107328 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107340 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107353 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107365 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107376 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107388 4809 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107401 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107414 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107427 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107438 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107449 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107461 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107472 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107484 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107502 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107512 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107525 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107537 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107548 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107562 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107574 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107586 4809 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107597 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107608 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107621 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107650 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107663 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107675 4809 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107686 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107697 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107708 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107722 4809 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107736 4809 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107747 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107763 4809 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107775 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107787 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107798 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107810 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107822 4809 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.107834 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.106021 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.115348 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.120667 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.133131 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.147815 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.153926 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.161076 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.176410 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.187483 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.198564 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.208934 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.209448 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.221285 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.232225 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.232682 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.241453 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.246314 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.248351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 11:08:44 crc kubenswrapper[4809]: W1205 11:08:44.248984 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-4f108168d1804a7537f92eae2ea97239a2c5ebae1d11e52d54693479ceb3927f WatchSource:0}: Error finding container 4f108168d1804a7537f92eae2ea97239a2c5ebae1d11e52d54693479ceb3927f: Status 404 returned error can't find the container with id 4f108168d1804a7537f92eae2ea97239a2c5ebae1d11e52d54693479ceb3927f Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.260140 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.271896 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.280891 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.289352 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: W1205 11:08:44.307188 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-16f3ff1d58a2a210c84abf32f4f607e76f4f9b519aebfdf1eca1a588e1949483 WatchSource:0}: Error finding container 16f3ff1d58a2a210c84abf32f4f607e76f4f9b519aebfdf1eca1a588e1949483: Status 404 returned error can't find the container with id 16f3ff1d58a2a210c84abf32f4f607e76f4f9b519aebfdf1eca1a588e1949483 Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.512661 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.512842 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:08:45.512822521 +0000 UTC m=+20.903799079 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.615451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.615503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.615531 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.615559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.615746 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.615766 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.615779 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.615841 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:45.615822671 +0000 UTC m=+21.006799229 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616727 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616769 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:45.616756186 +0000 UTC m=+21.007732744 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616821 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616848 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:45.616840068 +0000 UTC m=+21.007816626 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616897 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616908 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616918 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: E1205 11:08:44.616947 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:45.616937601 +0000 UTC m=+21.007914159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.876298 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.876953 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.878313 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.879026 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.880133 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.880829 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.881731 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.882806 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.883489 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.884610 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.885182 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.886406 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.887013 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.887656 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.888692 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.889289 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.890551 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.891120 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.891450 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.891844 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.892976 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.893486 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.894758 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.895237 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.896427 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.896944 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.897624 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.899338 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.899891 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.900995 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.901514 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.902484 4809 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.902594 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.904478 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.905517 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.906003 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.907435 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.908001 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.908803 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.909805 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.910498 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.911906 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.912411 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.913538 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.914227 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.915309 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.915850 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.917090 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.917712 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.919004 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.919533 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.920030 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.920502 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.921197 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.922295 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.923005 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.923613 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.938243 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.948462 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.958378 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.968267 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.978557 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:44 crc kubenswrapper[4809]: I1205 11:08:44.987705 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.029469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e6283e5a5b8b3fc87a44b33a77313b8b66a0608f61b157033bdc5dc061ad552f"} Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.031511 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a"} Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.031551 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76"} Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.031564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4f108168d1804a7537f92eae2ea97239a2c5ebae1d11e52d54693479ceb3927f"} Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.033980 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d"} Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.034038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"16f3ff1d58a2a210c84abf32f4f607e76f4f9b519aebfdf1eca1a588e1949483"} Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.046946 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.066356 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.080589 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.103218 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.115852 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.132454 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.147264 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.163990 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.181652 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.197839 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.212699 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.227028 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.242445 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.257373 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.275900 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.300071 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.316602 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.336907 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.528348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.528564 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:08:47.528530994 +0000 UTC m=+22.919507572 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.630040 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.630182 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.630520 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:47.630499266 +0000 UTC m=+23.021475884 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.630415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.630721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.630792 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.630995 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.630997 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631013 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631035 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631036 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631062 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631065 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:47.631055671 +0000 UTC m=+23.022032299 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631180 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:47.631151823 +0000 UTC m=+23.022128421 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631711 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.631762 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:47.63175232 +0000 UTC m=+23.022728958 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.871553 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.871993 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:45 crc kubenswrapper[4809]: I1205 11:08:45.871942 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.872262 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.872744 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:45 crc kubenswrapper[4809]: E1205 11:08:45.872971 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.040359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409"} Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.053915 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.074188 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.087731 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.102439 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.117360 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.133485 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.147354 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.161091 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.172469 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.547370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.547507 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:08:51.547484765 +0000 UTC m=+26.938461333 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.648196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.648298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648330 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.648366 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648419 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:51.648394958 +0000 UTC m=+27.039371556 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.648452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648529 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648544 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648565 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648587 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648591 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648611 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648613 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648622 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:51.648596854 +0000 UTC m=+27.039573452 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648793 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:51.648761118 +0000 UTC m=+27.039737716 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.648824 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:51.648809009 +0000 UTC m=+27.039785617 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.729883 4809 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.732384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.732449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.732467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.732550 4809 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.745376 4809 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.745851 4809 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.747211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.747364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.747388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.747415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.747433 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:47Z","lastTransitionTime":"2025-12-05T11:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.771083 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.777010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.777106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.777133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.777175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.777200 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:47Z","lastTransitionTime":"2025-12-05T11:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.801655 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.807609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.807669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.807682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.807702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.807714 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:47Z","lastTransitionTime":"2025-12-05T11:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.836711 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.843977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.844034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.844052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.844076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.844094 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:47Z","lastTransitionTime":"2025-12-05T11:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.867521 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.871513 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.871535 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.871575 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.871708 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.871831 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.872045 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.874492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.874597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.874623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.874707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.874730 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:47Z","lastTransitionTime":"2025-12-05T11:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.898262 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:47Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:47 crc kubenswrapper[4809]: E1205 11:08:47.898397 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.900556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.900596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.900608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.900975 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:47 crc kubenswrapper[4809]: I1205 11:08:47.901013 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:47Z","lastTransitionTime":"2025-12-05T11:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.004854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.004915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.004927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.004948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.004963 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.108204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.108251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.108263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.108284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.108294 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.210598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.210646 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.210655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.210669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.210678 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.313027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.313063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.313071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.313087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.313098 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.415876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.415917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.415926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.415940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.415949 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.518622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.518694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.518708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.518730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.518750 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.621768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.621817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.621828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.621847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.621859 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.724407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.724473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.724490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.724513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.724530 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.827304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.827360 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.827371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.827389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.827401 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.943076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.943107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.943115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.943130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:48 crc kubenswrapper[4809]: I1205 11:08:48.943139 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:48Z","lastTransitionTime":"2025-12-05T11:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.045139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.045224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.045239 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.045254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.045266 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.148048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.148087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.148097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.148113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.148122 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.246153 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zwhnj"] Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.246870 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.258498 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-s4hls"] Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.258959 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.274146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.274180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.274189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.274204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.274214 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.276110 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-n79nb"] Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.276398 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: W1205 11:08:49.277217 4809 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 05 11:08:49 crc kubenswrapper[4809]: E1205 11:08:49.277257 4809 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 11:08:49 crc kubenswrapper[4809]: W1205 11:08:49.278949 4809 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 05 11:08:49 crc kubenswrapper[4809]: E1205 11:08:49.278978 4809 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 11:08:49 crc kubenswrapper[4809]: W1205 11:08:49.314017 4809 reflector.go:561] object-"openshift-machine-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.314085 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 11:08:49 crc kubenswrapper[4809]: E1205 11:08:49.314109 4809 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.314316 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.314334 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.314490 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.315440 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.315972 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-rjsd4"] Dec 05 11:08:49 crc kubenswrapper[4809]: W1205 11:08:49.276850 4809 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 05 11:08:49 crc kubenswrapper[4809]: E1205 11:08:49.316307 4809 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.316324 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.320023 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.320283 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.328445 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.328664 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.328774 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.332163 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.372990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38775c19-63c3-49a0-974e-5629199dcd55-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373043 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwrpb\" (UniqueName: \"kubernetes.io/projected/38775c19-63c3-49a0-974e-5629199dcd55-kube-api-access-dwrpb\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373120 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87517ccc-4f19-430e-9d03-d71da4cbe196-proxy-tls\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373161 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-system-cni-dir\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373179 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-os-release\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373465 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38775c19-63c3-49a0-974e-5629199dcd55-cni-binary-copy\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373505 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbtl8\" (UniqueName: \"kubernetes.io/projected/87517ccc-4f19-430e-9d03-d71da4cbe196-kube-api-access-bbtl8\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-cnibin\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373676 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87517ccc-4f19-430e-9d03-d71da4cbe196-mcd-auth-proxy-config\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.373703 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/87517ccc-4f19-430e-9d03-d71da4cbe196-rootfs\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.376205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.376247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.376261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.376277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.376287 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.386085 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.400682 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.416907 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.433190 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.446684 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.465792 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-cnibin\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87517ccc-4f19-430e-9d03-d71da4cbe196-mcd-auth-proxy-config\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-netns\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475346 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/87517ccc-4f19-430e-9d03-d71da4cbe196-rootfs\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475372 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-daemon-config\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwrpb\" (UniqueName: \"kubernetes.io/projected/38775c19-63c3-49a0-974e-5629199dcd55-kube-api-access-dwrpb\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475401 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-cnibin\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475421 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87517ccc-4f19-430e-9d03-d71da4cbe196-proxy-tls\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475513 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-os-release\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475549 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38775c19-63c3-49a0-974e-5629199dcd55-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475534 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/87517ccc-4f19-430e-9d03-d71da4cbe196-rootfs\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-k8s-cni-cncf-io\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475806 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-cni-bin\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-system-cni-dir\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475947 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.475969 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-cnibin\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-system-cni-dir\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476018 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-os-release\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476058 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-conf-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476110 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-os-release\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476101 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbtl8\" (UniqueName: \"kubernetes.io/projected/87517ccc-4f19-430e-9d03-d71da4cbe196-kube-api-access-bbtl8\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7d9fdbc-ab41-4867-8e71-807ff609fb1b-hosts-file\") pod \"node-resolver-rjsd4\" (UID: \"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\") " pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476221 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-cni-binary-copy\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-system-cni-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476244 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38775c19-63c3-49a0-974e-5629199dcd55-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-cni-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476314 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-cni-multus\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476333 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38775c19-63c3-49a0-974e-5629199dcd55-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476354 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-multus-certs\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476380 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-socket-dir-parent\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476415 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-kubelet\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476435 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-hostroot\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476479 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfqw5\" (UniqueName: \"kubernetes.io/projected/d7d9fdbc-ab41-4867-8e71-807ff609fb1b-kube-api-access-pfqw5\") pod \"node-resolver-rjsd4\" (UID: \"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\") " pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38775c19-63c3-49a0-974e-5629199dcd55-cni-binary-copy\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476568 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-etc-kubernetes\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgjfn\" (UniqueName: \"kubernetes.io/projected/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-kube-api-access-jgjfn\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.476997 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38775c19-63c3-49a0-974e-5629199dcd55-cni-binary-copy\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.480743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.480822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.480847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.480879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.480901 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.482719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/87517ccc-4f19-430e-9d03-d71da4cbe196-proxy-tls\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.490350 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.513493 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwrpb\" (UniqueName: \"kubernetes.io/projected/38775c19-63c3-49a0-974e-5629199dcd55-kube-api-access-dwrpb\") pod \"multus-additional-cni-plugins-zwhnj\" (UID: \"38775c19-63c3-49a0-974e-5629199dcd55\") " pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.517563 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.535744 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.552162 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.568741 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577078 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-daemon-config\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577139 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-os-release\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-k8s-cni-cncf-io\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-cni-bin\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-cnibin\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-conf-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7d9fdbc-ab41-4867-8e71-807ff609fb1b-hosts-file\") pod \"node-resolver-rjsd4\" (UID: \"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\") " pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-os-release\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577285 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-k8s-cni-cncf-io\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-cni-binary-copy\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577399 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-cni-bin\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577400 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-system-cni-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577448 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-cni-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577471 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-cni-multus\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-multus-certs\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577524 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-socket-dir-parent\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577549 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-kubelet\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577556 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-cni-multus\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-hostroot\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfqw5\" (UniqueName: \"kubernetes.io/projected/d7d9fdbc-ab41-4867-8e71-807ff609fb1b-kube-api-access-pfqw5\") pod \"node-resolver-rjsd4\" (UID: \"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\") " pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577526 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-system-cni-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-etc-kubernetes\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-var-lib-kubelet\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577360 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-cnibin\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-conf-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-etc-kubernetes\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d7d9fdbc-ab41-4867-8e71-807ff609fb1b-hosts-file\") pod \"node-resolver-rjsd4\" (UID: \"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\") " pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577834 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-multus-certs\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577856 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgjfn\" (UniqueName: \"kubernetes.io/projected/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-kube-api-access-jgjfn\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577883 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-hostroot\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.577975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-socket-dir-parent\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.578045 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-netns\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.578141 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-host-run-netns\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.578130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-cni-dir\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.578439 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-multus-daemon-config\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.578460 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-cni-binary-copy\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.583599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.583660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.583671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.583692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.583707 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.584364 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.594763 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgjfn\" (UniqueName: \"kubernetes.io/projected/238d6f1b-f3b4-4b33-a33c-10a0c99e8178-kube-api-access-jgjfn\") pod \"multus-n79nb\" (UID: \"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\") " pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.599530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfqw5\" (UniqueName: \"kubernetes.io/projected/d7d9fdbc-ab41-4867-8e71-807ff609fb1b-kube-api-access-pfqw5\") pod \"node-resolver-rjsd4\" (UID: \"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\") " pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.605664 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.633782 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.647077 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-n79nb" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.650144 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: W1205 11:08:49.656760 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod238d6f1b_f3b4_4b33_a33c_10a0c99e8178.slice/crio-d646a238bad3ac093aa608d19cc7b885bfa74d2b78f765f64beb9bf2515ee7f4 WatchSource:0}: Error finding container d646a238bad3ac093aa608d19cc7b885bfa74d2b78f765f64beb9bf2515ee7f4: Status 404 returned error can't find the container with id d646a238bad3ac093aa608d19cc7b885bfa74d2b78f765f64beb9bf2515ee7f4 Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.656936 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rjsd4" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.674405 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.688061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.688098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.688111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.688128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.688140 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.692238 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.712701 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.714306 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cdvk8"] Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.715255 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.717036 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.717713 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.718389 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.718545 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.718670 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.718784 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.718880 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.734465 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.752167 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.767216 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.777508 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.790850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.790891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.790904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.790924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.790936 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.796742 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.815500 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.844756 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.862910 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.871534 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:49 crc kubenswrapper[4809]: E1205 11:08:49.871693 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.872111 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:49 crc kubenswrapper[4809]: E1205 11:08:49.872182 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.872241 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:49 crc kubenswrapper[4809]: E1205 11:08:49.872310 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.883918 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-bin\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.883994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-kubelet\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884016 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-log-socket\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-script-lib\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884180 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-slash\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884236 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-systemd-units\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-env-overrides\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884295 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-ovn-kubernetes\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884326 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-systemd\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884352 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-ovn\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884375 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-node-log\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884450 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4kb5\" (UniqueName: \"kubernetes.io/projected/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-kube-api-access-j4kb5\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884512 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-etc-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884554 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-netd\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884581 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-config\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884610 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-var-lib-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884658 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-netns\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884796 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.884818 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovn-node-metrics-cert\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.889102 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.893820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.893879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.893892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.893914 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.893928 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:49Z","lastTransitionTime":"2025-12-05T11:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.930556 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999027 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-slash\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-systemd-units\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999113 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-env-overrides\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-ovn-kubernetes\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-systemd\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999168 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-ovn\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-node-log\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4kb5\" (UniqueName: \"kubernetes.io/projected/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-kube-api-access-j4kb5\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-etc-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-netd\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999311 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-config\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999328 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-var-lib-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-netns\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999409 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovn-node-metrics-cert\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999464 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-kubelet\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-log-socket\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-bin\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-script-lib\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999820 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-netd\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999893 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-slash\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:49 crc kubenswrapper[4809]: I1205 11:08:49.999934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-systemd-units\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.000563 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-env-overrides\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.000621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-ovn-kubernetes\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.000679 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-systemd\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.000715 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-ovn\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.000751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-node-log\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.000915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-script-lib\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001051 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-etc-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001231 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001485 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-config\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001543 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-var-lib-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001576 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-kubelet\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001604 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-log-socket\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-bin\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001738 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-openvswitch\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001782 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-netns\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.001818 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.005774 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.005857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovn-node-metrics-cert\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.049697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rjsd4" event={"ID":"d7d9fdbc-ab41-4867-8e71-807ff609fb1b","Type":"ContainerStarted","Data":"845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.049750 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rjsd4" event={"ID":"d7d9fdbc-ab41-4867-8e71-807ff609fb1b","Type":"ContainerStarted","Data":"29bd8caa4ad813b742f291eb706cfc6a84fc6d1495c5d94fe4f172e84345fd62"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.050880 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.051618 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerStarted","Data":"ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.051664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerStarted","Data":"d646a238bad3ac093aa608d19cc7b885bfa74d2b78f765f64beb9bf2515ee7f4"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.065133 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4kb5\" (UniqueName: \"kubernetes.io/projected/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-kube-api-access-j4kb5\") pod \"ovnkube-node-cdvk8\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.103875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.103909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.103920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.103963 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.103975 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.131793 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.145652 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.165311 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.180254 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.192924 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.205482 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.205813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.205838 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.205846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.205858 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.205868 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.217429 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.229006 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.241943 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.253059 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.272352 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.272625 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.283590 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.286602 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.294955 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.297026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/87517ccc-4f19-430e-9d03-d71da4cbe196-mcd-auth-proxy-config\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.305520 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.309523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.309583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.309596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.309618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.309657 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.325317 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.338610 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.341425 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:50 crc kubenswrapper[4809]: W1205 11:08:50.351464 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf22f1c5_aea5_4717_b2b0_3d6a577ea53f.slice/crio-69838862339e86709318f40582d66ac51e9d319f9ca09bc0a40b49034ea0cc87 WatchSource:0}: Error finding container 69838862339e86709318f40582d66ac51e9d319f9ca09bc0a40b49034ea0cc87: Status 404 returned error can't find the container with id 69838862339e86709318f40582d66ac51e9d319f9ca09bc0a40b49034ea0cc87 Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.354680 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.366479 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.382537 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.395911 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.410833 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.415150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.415204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.415220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.415244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.415264 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: E1205 11:08:50.492502 4809 projected.go:288] Couldn't get configMap openshift-machine-config-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 05 11:08:50 crc kubenswrapper[4809]: E1205 11:08:50.492742 4809 projected.go:194] Error preparing data for projected volume kube-api-access-bbtl8 for pod openshift-machine-config-operator/machine-config-daemon-s4hls: failed to sync configmap cache: timed out waiting for the condition Dec 05 11:08:50 crc kubenswrapper[4809]: E1205 11:08:50.492868 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/87517ccc-4f19-430e-9d03-d71da4cbe196-kube-api-access-bbtl8 podName:87517ccc-4f19-430e-9d03-d71da4cbe196 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:50.992848676 +0000 UTC m=+26.383825224 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-bbtl8" (UniqueName: "kubernetes.io/projected/87517ccc-4f19-430e-9d03-d71da4cbe196-kube-api-access-bbtl8") pod "machine-config-daemon-s4hls" (UID: "87517ccc-4f19-430e-9d03-d71da4cbe196") : failed to sync configmap cache: timed out waiting for the condition Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.517572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.517832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.517895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.517959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.518021 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.617281 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-multus/multus-additional-cni-plugins-zwhnj" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.617698 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.620104 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.620146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.620159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.620175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.620188 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.699973 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.722917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.723150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.723271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.723362 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.723457 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.826366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.826602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.826684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.826752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.826820 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.886501 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.929224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.929269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.929281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.929298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:50 crc kubenswrapper[4809]: I1205 11:08:50.929312 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:50Z","lastTransitionTime":"2025-12-05T11:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.011923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbtl8\" (UniqueName: \"kubernetes.io/projected/87517ccc-4f19-430e-9d03-d71da4cbe196-kube-api-access-bbtl8\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.022068 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbtl8\" (UniqueName: \"kubernetes.io/projected/87517ccc-4f19-430e-9d03-d71da4cbe196-kube-api-access-bbtl8\") pod \"machine-config-daemon-s4hls\" (UID: \"87517ccc-4f19-430e-9d03-d71da4cbe196\") " pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.031250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.031407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.031503 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.031596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.031714 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.056511 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f" exitCode=0 Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.056592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.056907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"69838862339e86709318f40582d66ac51e9d319f9ca09bc0a40b49034ea0cc87"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.070116 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.090693 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.112707 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.130418 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.132436 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.134309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.134335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.134343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.134357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.134366 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.149599 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.163292 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.178171 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.192667 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.208379 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.220660 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.235618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.237092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.237121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.237132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.237147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.237159 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.251565 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.266966 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.281846 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:51Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.340026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.340087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.340098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.340122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.340146 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: W1205 11:08:51.405718 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38775c19_63c3_49a0_974e_5629199dcd55.slice/crio-596bf8a66b5f508fd0e11e1ffaf60a4f4155c92d2a7c628381de831233484e54 WatchSource:0}: Error finding container 596bf8a66b5f508fd0e11e1ffaf60a4f4155c92d2a7c628381de831233484e54: Status 404 returned error can't find the container with id 596bf8a66b5f508fd0e11e1ffaf60a4f4155c92d2a7c628381de831233484e54 Dec 05 11:08:51 crc kubenswrapper[4809]: W1205 11:08:51.406398 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87517ccc_4f19_430e_9d03_d71da4cbe196.slice/crio-e54d3d41895bd9f794614731e80e3b5292bb5f6b92431584972f99333711ec8d WatchSource:0}: Error finding container e54d3d41895bd9f794614731e80e3b5292bb5f6b92431584972f99333711ec8d: Status 404 returned error can't find the container with id e54d3d41895bd9f794614731e80e3b5292bb5f6b92431584972f99333711ec8d Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.442253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.442326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.442345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.442368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.442420 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.546059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.546090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.546098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.546113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.546122 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.621197 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.621359 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:08:59.62133457 +0000 UTC m=+35.012311128 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.649335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.649380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.649392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.649409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.649419 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.722386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.722431 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.722451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.722479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722574 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722618 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:59.722605214 +0000 UTC m=+35.113581772 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722697 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722722 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:59.722716097 +0000 UTC m=+35.113692655 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722784 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722798 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722810 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722831 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:59.72282537 +0000 UTC m=+35.113801928 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722871 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722881 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722888 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.722907 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:08:59.722901932 +0000 UTC m=+35.113878490 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.751937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.751963 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.751971 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.751983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.751993 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.854852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.854887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.854899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.854915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.854927 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.871381 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.871500 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.871834 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.871884 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.871922 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:51 crc kubenswrapper[4809]: E1205 11:08:51.871957 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.957247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.957285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.957297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.957312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:51 crc kubenswrapper[4809]: I1205 11:08:51.957324 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:51Z","lastTransitionTime":"2025-12-05T11:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.063685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.063747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.063763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.063786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.063797 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.066086 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.066120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.066134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.066937 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerStarted","Data":"b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.066960 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerStarted","Data":"596bf8a66b5f508fd0e11e1ffaf60a4f4155c92d2a7c628381de831233484e54"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.069498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.069532 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.069545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"e54d3d41895bd9f794614731e80e3b5292bb5f6b92431584972f99333711ec8d"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.080605 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.096216 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.116441 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.133885 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.146806 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.157737 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.169800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.169833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.169842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.169857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.169870 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.176074 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.187710 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.203499 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.220796 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.236080 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.250351 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.262768 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.271219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.271244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.271253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.297143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.297193 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.300352 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.323237 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.347989 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.358391 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.396925 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.463092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.463129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.463140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.463156 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.463167 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.467225 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.515065 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.544976 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.561393 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.566023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.566454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.566580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.566716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.566932 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.577125 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.597324 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.612549 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.632755 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.651058 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.669012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.669061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.669076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.669091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.669105 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.672287 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.772029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.772233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.772328 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.772410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.772497 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.874819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.875035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.875093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.875168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.875225 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.977835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.977965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.977984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.978005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:52 crc kubenswrapper[4809]: I1205 11:08:52.978021 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:52Z","lastTransitionTime":"2025-12-05T11:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.080271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.080546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.080665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.080747 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.080807 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.087142 4809 generic.go:334] "Generic (PLEG): container finished" podID="38775c19-63c3-49a0-974e-5629199dcd55" containerID="b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505" exitCode=0 Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.087235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerDied","Data":"b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.090476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.090519 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.090532 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.123823 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.146577 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.161331 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-l5kgl"] Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.161714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.163695 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.165552 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.165647 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.174796 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.175058 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.182888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.182913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.182922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.182934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.182943 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.187938 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.209618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.223902 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.235221 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.249004 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.260277 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhmm6\" (UniqueName: \"kubernetes.io/projected/0eb95f25-779a-4461-82d1-b7ea053eb2c3-kube-api-access-jhmm6\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.260341 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0eb95f25-779a-4461-82d1-b7ea053eb2c3-serviceca\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.260366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0eb95f25-779a-4461-82d1-b7ea053eb2c3-host\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.266528 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.277757 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.285233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.285267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.285278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.285293 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.285306 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.288185 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.303265 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.325368 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.355327 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.361261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0eb95f25-779a-4461-82d1-b7ea053eb2c3-host\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.361507 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhmm6\" (UniqueName: \"kubernetes.io/projected/0eb95f25-779a-4461-82d1-b7ea053eb2c3-kube-api-access-jhmm6\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.361592 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0eb95f25-779a-4461-82d1-b7ea053eb2c3-serviceca\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.361375 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0eb95f25-779a-4461-82d1-b7ea053eb2c3-host\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.362561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0eb95f25-779a-4461-82d1-b7ea053eb2c3-serviceca\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.388623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.388690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.388706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.388725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.388739 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.393363 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhmm6\" (UniqueName: \"kubernetes.io/projected/0eb95f25-779a-4461-82d1-b7ea053eb2c3-kube-api-access-jhmm6\") pod \"node-ca-l5kgl\" (UID: \"0eb95f25-779a-4461-82d1-b7ea053eb2c3\") " pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.396700 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.416872 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.431425 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.444081 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.461380 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.471922 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.481751 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-l5kgl" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.484681 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.490619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.490654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.490664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.490680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.490691 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: W1205 11:08:53.496050 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0eb95f25_779a_4461_82d1_b7ea053eb2c3.slice/crio-8b75549893810634ad3341df0cac2f8f9fc57577e67381276c2a580dac6f2337 WatchSource:0}: Error finding container 8b75549893810634ad3341df0cac2f8f9fc57577e67381276c2a580dac6f2337: Status 404 returned error can't find the container with id 8b75549893810634ad3341df0cac2f8f9fc57577e67381276c2a580dac6f2337 Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.501937 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.522376 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.539694 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.551751 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.565810 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.600511 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.603590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.603615 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.603667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.603683 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.603693 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.614295 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.628697 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:53Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.707344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.707376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.707388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.707407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.707420 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.810428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.810464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.810476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.810492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.810503 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.871982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:53 crc kubenswrapper[4809]: E1205 11:08:53.872117 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.872184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:53 crc kubenswrapper[4809]: E1205 11:08:53.872244 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.872305 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:53 crc kubenswrapper[4809]: E1205 11:08:53.872361 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.914836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.914867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.914875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.914888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:53 crc kubenswrapper[4809]: I1205 11:08:53.914898 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:53Z","lastTransitionTime":"2025-12-05T11:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.018254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.018474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.018484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.018500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.018511 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.098053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerStarted","Data":"dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.099573 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-l5kgl" event={"ID":"0eb95f25-779a-4461-82d1-b7ea053eb2c3","Type":"ContainerStarted","Data":"0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.099647 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-l5kgl" event={"ID":"0eb95f25-779a-4461-82d1-b7ea053eb2c3","Type":"ContainerStarted","Data":"8b75549893810634ad3341df0cac2f8f9fc57577e67381276c2a580dac6f2337"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.121192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.121246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.121255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.121275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.121288 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.126976 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.141039 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.206223 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.222821 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.223698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.223755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.223768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.223790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.223812 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.237726 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.255717 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.273402 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.288187 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.314323 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.327580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.327649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.327662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.327680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.327694 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.328760 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.346116 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.363615 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.380775 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.399264 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.416038 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.428888 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.430332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.430380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.430392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.430411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.430421 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.441215 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.452203 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.460116 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.476250 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.486214 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.507254 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.518598 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.532815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.533060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.533145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.533241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.533313 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.534622 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.545448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.557144 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.571862 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.588906 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.600037 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.612104 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.635805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.635854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.635866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.635885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.635896 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.737931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.737979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.737990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.738007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.738020 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.840076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.840120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.840132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.840149 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.840160 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.885956 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.897573 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.912309 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.927830 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.943051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.943087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.943096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.943109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.943118 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:54Z","lastTransitionTime":"2025-12-05T11:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.953890 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.967942 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.982137 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:54 crc kubenswrapper[4809]: I1205 11:08:54.994162 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.007343 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.024102 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.044413 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.045721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.045763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.045772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.045786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.045796 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.054485 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.077143 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.091172 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.108837 4809 generic.go:334] "Generic (PLEG): container finished" podID="38775c19-63c3-49a0-974e-5629199dcd55" containerID="dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d" exitCode=0 Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.108947 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerDied","Data":"dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.116785 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.129396 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.145655 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.148139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.148185 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.148194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.148212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.148223 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.160450 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.182417 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.196127 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.210451 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.223605 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.245349 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.251116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.251165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.251180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.251200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.251212 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.260308 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.275614 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.289142 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.300264 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.312691 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.322982 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.333018 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.343046 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.353953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.353993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.354004 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.354036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.354049 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.456486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.456525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.456535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.456551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.456563 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.558772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.558812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.558822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.558837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.558850 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.661758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.661797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.661806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.661822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.661830 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.808252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.808299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.808309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.808327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.808339 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.871169 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.871250 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.871267 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:55 crc kubenswrapper[4809]: E1205 11:08:55.871333 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:55 crc kubenswrapper[4809]: E1205 11:08:55.871492 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:55 crc kubenswrapper[4809]: E1205 11:08:55.871691 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.910866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.910925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.910948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.910977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:55 crc kubenswrapper[4809]: I1205 11:08:55.911000 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:55Z","lastTransitionTime":"2025-12-05T11:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.013392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.013438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.013447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.013463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.013473 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.115715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.115771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.115789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.115811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.115827 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.218616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.218681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.218694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.218716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.218729 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.320573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.320657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.320675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.320700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.320719 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.422872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.423107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.423117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.423131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.423141 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.526452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.526511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.526546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.526569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.526584 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.643802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.643849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.643860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.643880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.643892 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.757420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.757533 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.757547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.757573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.757592 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.860017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.860056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.860068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.860083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.860093 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.963201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.963256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.963269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.963289 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:56 crc kubenswrapper[4809]: I1205 11:08:56.963303 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:56Z","lastTransitionTime":"2025-12-05T11:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.065463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.065534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.065556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.065577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.065589 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.137836 4809 generic.go:334] "Generic (PLEG): container finished" podID="38775c19-63c3-49a0-974e-5629199dcd55" containerID="cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707" exitCode=0 Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.137879 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerDied","Data":"cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.158842 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.170964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.171040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.171050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.171066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.171076 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.174037 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.188327 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.206781 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.222530 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.234520 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.248249 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.261322 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.273737 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.274118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.274152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.274164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.274180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.274190 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.332363 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.347505 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.358491 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.368405 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.376562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.376606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.376616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.376648 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.376663 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.384548 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.394099 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:57Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.479529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.479567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.479579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.479595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.479607 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.582589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.582662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.582675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.582691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.582702 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.685212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.685266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.685278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.685300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.685313 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.787832 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.787866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.787876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.787895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.787905 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.871063 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.871139 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.871177 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:57 crc kubenswrapper[4809]: E1205 11:08:57.871326 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:57 crc kubenswrapper[4809]: E1205 11:08:57.871452 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:57 crc kubenswrapper[4809]: E1205 11:08:57.871575 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.890009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.890051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.890068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.890086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.890101 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.992610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.992674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.992685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.992701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:57 crc kubenswrapper[4809]: I1205 11:08:57.992713 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:57Z","lastTransitionTime":"2025-12-05T11:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.026977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.027180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.027325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.027523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.027691 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: E1205 11:08:58.040096 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.044123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.044252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.044330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.044412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.044491 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: E1205 11:08:58.057039 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.059778 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.059812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.059823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.059838 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.059849 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: E1205 11:08:58.071230 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.074224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.074255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.074263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.074307 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.074316 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: E1205 11:08:58.087682 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.091128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.091245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.091320 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.091391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.091448 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: E1205 11:08:58.103961 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: E1205 11:08:58.104221 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.105707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.105825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.105933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.106036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.106117 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.144532 4809 generic.go:334] "Generic (PLEG): container finished" podID="38775c19-63c3-49a0-974e-5629199dcd55" containerID="b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e" exitCode=0 Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.144597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerDied","Data":"b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.151388 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.152708 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.152736 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.165758 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.181456 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.191912 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.194571 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.196344 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.208408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.208440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.208449 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.208469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.208481 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.226168 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.239929 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.251877 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.262300 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.279456 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.290345 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.304023 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.314956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.315102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.315132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.315166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.315194 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.326423 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.344129 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.363679 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.374495 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.389745 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.402890 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.414786 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.419019 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.419043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.419051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.419064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.419075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.435497 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.458055 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.489872 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.503607 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.521298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.521340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.521349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.521364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.521374 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.523924 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.535762 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.556043 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.572270 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.584056 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.595540 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.610793 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.625607 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.627267 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.627294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.627304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.627319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.627328 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.649309 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:58Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.730209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.730249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.730259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.730274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.730283 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.833169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.833196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.833207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.833224 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.833240 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.935656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.935708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.935726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.935749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:58 crc kubenswrapper[4809]: I1205 11:08:58.935765 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:58Z","lastTransitionTime":"2025-12-05T11:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.038356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.038407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.038419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.038438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.038453 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.140940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.140986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.140997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.141013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.141024 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.158019 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerStarted","Data":"7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.158119 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.175187 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.190726 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.209467 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.225323 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.237725 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.243054 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.243093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.243103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.243117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.243128 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.298710 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.319834 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.338351 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.345716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.345758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.345766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.345779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.345787 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.351493 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.362826 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.389231 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.402304 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.420467 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.430212 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.447651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.447692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.447707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.447725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.447738 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.459766 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:08:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.550364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.550674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.550763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.550868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.550958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.654527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.654831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.654894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.654954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.655008 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.690341 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.690593 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:09:15.690574967 +0000 UTC m=+51.081551525 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.757854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.757884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.757891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.757903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.757912 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.791334 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.791391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.791419 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.791459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.791583 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.791658 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:15.791623714 +0000 UTC m=+51.182600272 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792030 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792079 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:15.792068316 +0000 UTC m=+51.183044874 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792147 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792162 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792176 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792207 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:15.792197729 +0000 UTC m=+51.183174287 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792262 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792277 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792288 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.792314 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:15.792306452 +0000 UTC m=+51.183283010 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.860974 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.861009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.861018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.861033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.861042 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.871374 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.871489 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.871787 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.871839 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.871879 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:08:59 crc kubenswrapper[4809]: E1205 11:08:59.871914 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.963552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.963760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.963926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.964089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:08:59 crc kubenswrapper[4809]: I1205 11:08:59.964184 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:08:59Z","lastTransitionTime":"2025-12-05T11:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.066724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.066765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.066774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.066790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.066800 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.161105 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.169130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.169310 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.169374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.169467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.169534 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.272229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.272282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.272296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.272314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.272329 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.374740 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.374774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.374785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.374799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.374808 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.477588 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.477985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.477999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.478016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.478031 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.581621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.581692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.581705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.581725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.581740 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.684948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.684994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.685003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.685018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.685027 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.788802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.788849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.788861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.788907 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.788924 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.891861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.891905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.891917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.891935 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.891946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.996095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.996164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.996187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.996216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:00 crc kubenswrapper[4809]: I1205 11:09:00.996249 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:00Z","lastTransitionTime":"2025-12-05T11:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.099297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.099353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.099371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.099395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.099413 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.167599 4809 generic.go:334] "Generic (PLEG): container finished" podID="38775c19-63c3-49a0-974e-5629199dcd55" containerID="7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce" exitCode=0 Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.167666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerDied","Data":"7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.192135 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.201938 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.201980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.201994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.202009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.202020 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.208942 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.220959 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.241525 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.254187 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.276189 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.289533 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.304929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.304970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.304982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.305002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.305015 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.305574 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.316216 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.328691 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.344385 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.357452 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.369337 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.383431 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.395619 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:01Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.407136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.407173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.407185 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.407202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.407212 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.509519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.509573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.509590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.509611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.509655 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.612380 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.612422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.612432 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.612447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.612457 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.716790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.716855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.716872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.716902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.716919 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.819085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.819131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.819146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.819161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.819171 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.871307 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.871352 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.871397 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:01 crc kubenswrapper[4809]: E1205 11:09:01.871466 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:01 crc kubenswrapper[4809]: E1205 11:09:01.871588 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:01 crc kubenswrapper[4809]: E1205 11:09:01.871731 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.922025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.922101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.922120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.922146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:01 crc kubenswrapper[4809]: I1205 11:09:01.922165 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:01Z","lastTransitionTime":"2025-12-05T11:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.025780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.025870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.025892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.025925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.025946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.128811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.128888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.128905 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.128930 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.128954 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.231954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.232018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.232035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.232059 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.232076 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.342750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.342807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.342876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.342903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.342921 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.445067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.445110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.445119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.445138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.445147 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.548038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.548074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.548085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.548102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.548114 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.651714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.651760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.651772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.651791 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.651806 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.755469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.755519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.755531 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.755551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.755564 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.857495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.857549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.857564 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.857585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.857614 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.961002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.961070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.961093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.961124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:02 crc kubenswrapper[4809]: I1205 11:09:02.961144 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:02Z","lastTransitionTime":"2025-12-05T11:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.053681 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv"] Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.054294 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.059100 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.059507 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.064322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.064377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.064399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.064429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.064452 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.096307 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.116302 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.128114 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.138781 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.155872 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.166600 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.168197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.168246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.168262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.168285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.168302 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.176250 4809 generic.go:334] "Generic (PLEG): container finished" podID="38775c19-63c3-49a0-974e-5629199dcd55" containerID="a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a" exitCode=0 Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.176320 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerDied","Data":"a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.176343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66d25413-5343-45fe-acb6-3d72fb5a4e3e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.176406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8h9z\" (UniqueName: \"kubernetes.io/projected/66d25413-5343-45fe-acb6-3d72fb5a4e3e-kube-api-access-g8h9z\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.176439 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66d25413-5343-45fe-acb6-3d72fb5a4e3e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.176508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66d25413-5343-45fe-acb6-3d72fb5a4e3e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.178490 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/0.log" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.183392 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.183758 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195" exitCode=1 Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.183816 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.184721 4809 scope.go:117] "RemoveContainer" containerID="bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.198034 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.211004 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.227679 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.242364 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.261611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.271500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.271542 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.271554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.271575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.271588 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.277145 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.277796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66d25413-5343-45fe-acb6-3d72fb5a4e3e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.277820 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66d25413-5343-45fe-acb6-3d72fb5a4e3e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.277905 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8h9z\" (UniqueName: \"kubernetes.io/projected/66d25413-5343-45fe-acb6-3d72fb5a4e3e-kube-api-access-g8h9z\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.277933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66d25413-5343-45fe-acb6-3d72fb5a4e3e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.280944 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/66d25413-5343-45fe-acb6-3d72fb5a4e3e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.281038 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/66d25413-5343-45fe-acb6-3d72fb5a4e3e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.287093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/66d25413-5343-45fe-acb6-3d72fb5a4e3e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.292900 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.298874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8h9z\" (UniqueName: \"kubernetes.io/projected/66d25413-5343-45fe-acb6-3d72fb5a4e3e-kube-api-access-g8h9z\") pod \"ovnkube-control-plane-749d76644c-ssnzv\" (UID: \"66d25413-5343-45fe-acb6-3d72fb5a4e3e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.306373 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.319448 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.341786 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.354209 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.365780 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.372423 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.373066 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.373901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.373927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.373939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.373955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.373965 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.388657 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 11:09:02.863380 6008 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 11:09:02.863404 6008 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 11:09:02.863420 6008 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 11:09:02.863425 6008 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 11:09:02.863441 6008 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 11:09:02.863492 6008 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 11:09:02.863536 6008 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 11:09:02.863562 6008 factory.go:656] Stopping watch factory\\\\nI1205 11:09:02.863582 6008 ovnkube.go:599] Stopped ovnkube\\\\nI1205 11:09:02.863675 6008 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 11:09:02.863730 6008 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 11:09:02.863741 6008 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 11:09:02.863749 6008 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 11:09:02.863757 6008 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 11:09:02.863765 6008 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 11:09:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.397224 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.409184 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.422105 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.435241 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.447736 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.460618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.476111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.476154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.476167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.476183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.476198 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.477097 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.489787 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.507672 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.518750 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.531952 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:03Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.578605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.578666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.578676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.578691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.578699 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.683125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.684339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.684750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.684890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.685062 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: W1205 11:09:03.700953 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66d25413_5343_45fe_acb6_3d72fb5a4e3e.slice/crio-c69b9ce09546823bcb397c2948dc440ab868f001ea8fa56030d562ebc4390521 WatchSource:0}: Error finding container c69b9ce09546823bcb397c2948dc440ab868f001ea8fa56030d562ebc4390521: Status 404 returned error can't find the container with id c69b9ce09546823bcb397c2948dc440ab868f001ea8fa56030d562ebc4390521 Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.788673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.788746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.788761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.788787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.788805 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.871661 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.871722 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:03 crc kubenswrapper[4809]: E1205 11:09:03.871814 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:03 crc kubenswrapper[4809]: E1205 11:09:03.871999 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.871674 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:03 crc kubenswrapper[4809]: E1205 11:09:03.872109 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.891188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.891246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.891256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.891277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.891289 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.989479 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xxkn4"] Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.990009 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:03 crc kubenswrapper[4809]: E1205 11:09:03.990084 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.994900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.994967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.994990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.995015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:03 crc kubenswrapper[4809]: I1205 11:09:03.995035 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:03Z","lastTransitionTime":"2025-12-05T11:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.010764 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.024301 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.038817 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.054282 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.067521 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.086081 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.086122 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skrvp\" (UniqueName: \"kubernetes.io/projected/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-kube-api-access-skrvp\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.094890 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 11:09:02.863380 6008 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 11:09:02.863404 6008 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 11:09:02.863420 6008 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 11:09:02.863425 6008 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 11:09:02.863441 6008 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 11:09:02.863492 6008 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 11:09:02.863536 6008 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 11:09:02.863562 6008 factory.go:656] Stopping watch factory\\\\nI1205 11:09:02.863582 6008 ovnkube.go:599] Stopped ovnkube\\\\nI1205 11:09:02.863675 6008 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 11:09:02.863730 6008 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 11:09:02.863741 6008 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 11:09:02.863749 6008 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 11:09:02.863757 6008 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 11:09:02.863765 6008 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 11:09:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.097357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.097378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.097394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.097407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.097419 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.109764 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.149617 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.174604 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.186433 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.186475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skrvp\" (UniqueName: \"kubernetes.io/projected/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-kube-api-access-skrvp\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:04 crc kubenswrapper[4809]: E1205 11:09:04.186618 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:04 crc kubenswrapper[4809]: E1205 11:09:04.186735 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:04.686716815 +0000 UTC m=+40.077693373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.192156 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" event={"ID":"38775c19-63c3-49a0-974e-5629199dcd55","Type":"ContainerStarted","Data":"39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.193621 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" event={"ID":"66d25413-5343-45fe-acb6-3d72fb5a4e3e","Type":"ContainerStarted","Data":"c69b9ce09546823bcb397c2948dc440ab868f001ea8fa56030d562ebc4390521"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.196851 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.197879 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/0.log" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.199706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.199750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.199768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.199788 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.199804 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.202163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.202317 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.214007 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skrvp\" (UniqueName: \"kubernetes.io/projected/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-kube-api-access-skrvp\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.226207 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.241247 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.255855 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.268719 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.281357 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.299108 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.301790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.301824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.301840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.301860 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.301874 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.357451 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.368585 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.380750 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.390680 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.401960 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.403750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.403777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.403787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.403802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.403812 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.413488 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.424613 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.434999 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.443313 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.451358 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.464387 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.479684 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.497308 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.505968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.506003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.506014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.506027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.506035 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.509744 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.524087 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.545584 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 11:09:02.863380 6008 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 11:09:02.863404 6008 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 11:09:02.863420 6008 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 11:09:02.863425 6008 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 11:09:02.863441 6008 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 11:09:02.863492 6008 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 11:09:02.863536 6008 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 11:09:02.863562 6008 factory.go:656] Stopping watch factory\\\\nI1205 11:09:02.863582 6008 ovnkube.go:599] Stopped ovnkube\\\\nI1205 11:09:02.863675 6008 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 11:09:02.863730 6008 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 11:09:02.863741 6008 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 11:09:02.863749 6008 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 11:09:02.863757 6008 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 11:09:02.863765 6008 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 11:09:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.557696 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.581384 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.608553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.608583 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.608595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.608610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.608620 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.693334 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:04 crc kubenswrapper[4809]: E1205 11:09:04.693486 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:04 crc kubenswrapper[4809]: E1205 11:09:04.693529 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:05.693515694 +0000 UTC m=+41.084492242 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.711017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.711055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.711067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.711083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.711097 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.812751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.812793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.812805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.812819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.812830 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.883379 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.898472 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.910835 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.914396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.914423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.914431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.914444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.914453 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:04Z","lastTransitionTime":"2025-12-05T11:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.921311 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.931484 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.950049 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.962149 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.974268 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:04 crc kubenswrapper[4809]: I1205 11:09:04.986544 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.005426 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 11:09:02.863380 6008 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 11:09:02.863404 6008 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 11:09:02.863420 6008 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 11:09:02.863425 6008 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 11:09:02.863441 6008 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 11:09:02.863492 6008 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 11:09:02.863536 6008 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 11:09:02.863562 6008 factory.go:656] Stopping watch factory\\\\nI1205 11:09:02.863582 6008 ovnkube.go:599] Stopped ovnkube\\\\nI1205 11:09:02.863675 6008 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 11:09:02.863730 6008 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 11:09:02.863741 6008 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 11:09:02.863749 6008 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 11:09:02.863757 6008 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 11:09:02.863765 6008 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 11:09:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.016136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.016161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.016168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.016181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.016190 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.017334 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.032299 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.044850 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.060263 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.071708 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.086166 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.104798 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.117729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.117772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.117785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.117802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.117815 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.148103 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf22f1c5_aea5_4717_b2b0_3d6a577ea53f.slice/crio-conmon-c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9.scope\": RecentStats: unable to find data in memory cache]" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.209115 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/1.log" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.210062 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/0.log" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.213851 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9" exitCode=1 Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.213937 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.213996 4809 scope.go:117] "RemoveContainer" containerID="bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.215342 4809 scope.go:117] "RemoveContainer" containerID="c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9" Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.215667 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.218190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" event={"ID":"66d25413-5343-45fe-acb6-3d72fb5a4e3e","Type":"ContainerStarted","Data":"c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.218524 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" event={"ID":"66d25413-5343-45fe-acb6-3d72fb5a4e3e","Type":"ContainerStarted","Data":"d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.219874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.219931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.219954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.219982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.220006 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.230735 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.245170 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.259675 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.272675 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.285879 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.304438 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.318907 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.330359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.330399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.330411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.330427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.330438 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.331210 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.343181 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.353946 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.369151 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.379779 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.398505 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 11:09:02.863380 6008 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 11:09:02.863404 6008 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 11:09:02.863420 6008 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 11:09:02.863425 6008 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 11:09:02.863441 6008 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 11:09:02.863492 6008 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 11:09:02.863536 6008 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 11:09:02.863562 6008 factory.go:656] Stopping watch factory\\\\nI1205 11:09:02.863582 6008 ovnkube.go:599] Stopped ovnkube\\\\nI1205 11:09:02.863675 6008 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 11:09:02.863730 6008 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 11:09:02.863741 6008 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 11:09:02.863749 6008 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 11:09:02.863757 6008 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 11:09:02.863765 6008 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 11:09:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"message\\\":\\\"k controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:04.983664 6237 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983674 6237 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1205 11:09:04.983682 6237 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1205 11:09:04.983688 6237 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983689 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-n79nb\\\\nI1205 11:09:04.983687 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-s4hls\\\\nI1205 11:09:04.983697 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1205 11:09:04.983703 6237 obj_retry.go:365] Adding new object\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.408817 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.431211 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.432622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.432902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.432937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.433003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.433023 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.445839 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.458209 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.474166 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.485188 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.498315 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.512860 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.526188 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.535994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.536039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.536055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.536080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.536097 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.541738 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.555869 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.567578 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.580410 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.595790 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.605176 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.624263 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.638096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.638135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.638147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.638164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.638177 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.664356 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.703421 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.703706 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.703805 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:07.70378195 +0000 UTC m=+43.094758598 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.712545 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb565562395f9c01997e4b8aeb06c82f05fbb2e95b42f4ddf7b92f893c3f0195\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 11:09:02.863380 6008 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 11:09:02.863404 6008 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 11:09:02.863420 6008 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 11:09:02.863425 6008 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 11:09:02.863441 6008 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 11:09:02.863492 6008 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1205 11:09:02.863536 6008 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 11:09:02.863562 6008 factory.go:656] Stopping watch factory\\\\nI1205 11:09:02.863582 6008 ovnkube.go:599] Stopped ovnkube\\\\nI1205 11:09:02.863675 6008 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 11:09:02.863730 6008 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 11:09:02.863741 6008 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 11:09:02.863749 6008 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 11:09:02.863757 6008 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 11:09:02.863765 6008 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 11:09:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"message\\\":\\\"k controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:04.983664 6237 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983674 6237 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1205 11:09:04.983682 6237 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1205 11:09:04.983688 6237 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983689 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-n79nb\\\\nI1205 11:09:04.983687 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-s4hls\\\\nI1205 11:09:04.983697 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1205 11:09:04.983703 6237 obj_retry.go:365] Adding new object\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.742729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.742780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.742794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.742813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.742825 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.749521 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.771398 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.805922 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.830321 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:05Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.845803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.845854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.845868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.845887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.845898 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.872124 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.872127 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.872213 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.872416 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.872489 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.872520 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.872662 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:05 crc kubenswrapper[4809]: E1205 11:09:05.872872 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.947951 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.947986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.947997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.948013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:05 crc kubenswrapper[4809]: I1205 11:09:05.948026 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:05Z","lastTransitionTime":"2025-12-05T11:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.050653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.050717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.050733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.050754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.050769 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.153714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.153768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.153782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.153800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.153811 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.224515 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/1.log" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.230401 4809 scope.go:117] "RemoveContainer" containerID="c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9" Dec 05 11:09:06 crc kubenswrapper[4809]: E1205 11:09:06.230687 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.257026 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.257121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.257142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.257165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.257183 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.264456 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.280589 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.296471 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.309731 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.341712 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"message\\\":\\\"k controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:04.983664 6237 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983674 6237 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1205 11:09:04.983682 6237 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1205 11:09:04.983688 6237 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983689 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-n79nb\\\\nI1205 11:09:04.983687 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-s4hls\\\\nI1205 11:09:04.983697 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1205 11:09:04.983703 6237 obj_retry.go:365] Adding new object\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.355542 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.362466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.362513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.362524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.362547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.362565 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.375537 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.392307 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.415122 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.427064 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.446136 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.459195 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.464992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.465029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.465045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.465061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.465071 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.472229 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.486391 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.497951 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.507573 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.519717 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:06Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.568064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.568108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.568118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.568136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.568147 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.670121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.670151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.670159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.670172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.670180 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.773434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.773537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.773563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.773593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.773615 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.876534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.876572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.876581 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.876595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.876605 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.979966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.980035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.980055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.980082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:06 crc kubenswrapper[4809]: I1205 11:09:06.980106 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:06Z","lastTransitionTime":"2025-12-05T11:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.082754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.082787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.082796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.082808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.082818 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.186029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.186092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.186105 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.186125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.186139 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.289255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.289354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.289391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.289420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.289446 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.392244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.392299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.392312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.392333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.392346 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.494919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.494984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.494996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.495023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.495037 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.597790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.597825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.597834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.597847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.597857 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.700139 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.700213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.700225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.700250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.700267 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.727682 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:07 crc kubenswrapper[4809]: E1205 11:09:07.727807 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:07 crc kubenswrapper[4809]: E1205 11:09:07.727870 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:11.727856317 +0000 UTC m=+47.118832875 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.803777 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.803841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.803859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.803886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.803905 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.871874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.871995 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:07 crc kubenswrapper[4809]: E1205 11:09:07.872103 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.872014 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:07 crc kubenswrapper[4809]: E1205 11:09:07.872182 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.872039 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:07 crc kubenswrapper[4809]: E1205 11:09:07.872302 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:07 crc kubenswrapper[4809]: E1205 11:09:07.872502 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.906496 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.907172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.907688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.907875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:07 crc kubenswrapper[4809]: I1205 11:09:07.907912 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:07Z","lastTransitionTime":"2025-12-05T11:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.010586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.010679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.010693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.010709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.010721 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.113336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.113375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.113388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.113406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.113419 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.203761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.203836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.203859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.203891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.203913 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: E1205 11:09:08.223253 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:08Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.227843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.227887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.227902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.227921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.227936 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: E1205 11:09:08.242288 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:08Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.246976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.247023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.247037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.247056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.247071 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: E1205 11:09:08.260521 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:08Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.264943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.264990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.265007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.265033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.265050 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: E1205 11:09:08.288515 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:08Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.295229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.295287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.295306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.295329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.295348 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: E1205 11:09:08.312990 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:08Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:08 crc kubenswrapper[4809]: E1205 11:09:08.313226 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.316820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.316937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.316962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.316992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.317014 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.420431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.420465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.420474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.420489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.420501 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.524308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.524389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.524416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.524476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.524495 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.628066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.628112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.628129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.628154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.628170 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.731702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.731768 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.731785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.731810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.731830 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.834715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.834754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.834763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.834775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.834783 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.937457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.937500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.937512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.937528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:08 crc kubenswrapper[4809]: I1205 11:09:08.937542 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:08Z","lastTransitionTime":"2025-12-05T11:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.040435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.040483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.040494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.040510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.040520 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.142852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.142910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.142922 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.142945 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.142958 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.246194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.246274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.246299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.246336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.246366 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.350025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.350067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.350077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.350097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.350108 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.454457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.454504 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.454516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.454530 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.454542 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.557080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.557123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.557131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.557145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.557156 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.659672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.659709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.659717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.659730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.659740 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.763879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.763936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.763956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.763986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.764007 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.866817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.866883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.866903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.866930 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.866947 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.871287 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:09 crc kubenswrapper[4809]: E1205 11:09:09.871403 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.871304 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:09 crc kubenswrapper[4809]: E1205 11:09:09.871465 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.871304 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:09 crc kubenswrapper[4809]: E1205 11:09:09.871560 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.871471 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:09 crc kubenswrapper[4809]: E1205 11:09:09.871619 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.969876 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.969910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.969918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.969932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:09 crc kubenswrapper[4809]: I1205 11:09:09.969944 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:09Z","lastTransitionTime":"2025-12-05T11:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.072347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.072390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.072400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.072413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.072422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.174996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.175056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.175069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.175083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.175093 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.277599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.277702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.277727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.277793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.277818 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.381327 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.381393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.381415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.381446 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.381468 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.485341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.485403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.485419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.485444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.485460 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.589124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.589177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.589195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.589217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.589234 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.692425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.692501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.692519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.692543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.692561 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.795750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.795819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.795838 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.795863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.795889 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.899246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.899306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.899322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.899348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:10 crc kubenswrapper[4809]: I1205 11:09:10.899365 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:10Z","lastTransitionTime":"2025-12-05T11:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.001744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.001789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.001798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.001812 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.001821 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.105052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.105096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.105106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.105121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.105132 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.207325 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.207364 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.207376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.207393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.207405 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.309508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.309544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.309553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.309566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.309575 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.412198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.412253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.412270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.412291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.412313 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.514925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.515036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.515049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.515066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.515078 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.617705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.617738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.617746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.617757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.617767 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.720682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.720730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.720744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.720765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.720781 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.769674 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:11 crc kubenswrapper[4809]: E1205 11:09:11.769875 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:11 crc kubenswrapper[4809]: E1205 11:09:11.770450 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:19.770426003 +0000 UTC m=+55.161402561 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.824098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.824129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.824154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.824169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.824178 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.871109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:11 crc kubenswrapper[4809]: E1205 11:09:11.871268 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.871542 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.871599 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.871550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:11 crc kubenswrapper[4809]: E1205 11:09:11.871750 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:11 crc kubenswrapper[4809]: E1205 11:09:11.871871 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:11 crc kubenswrapper[4809]: E1205 11:09:11.872103 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.926552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.926677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.926701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.926720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:11 crc kubenswrapper[4809]: I1205 11:09:11.926736 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:11Z","lastTransitionTime":"2025-12-05T11:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.030033 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.030100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.030115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.030136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.030149 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.133280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.133338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.133348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.133372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.133386 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.236677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.236842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.236872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.236904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.236929 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.340341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.340417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.340453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.340484 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.340511 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.444169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.444227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.444244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.444268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.444288 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.546541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.546574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.546582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.546594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.546602 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.649083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.649126 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.649135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.649154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.649164 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.752241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.752282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.752292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.752308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.752318 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.854202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.854235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.854243 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.854255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.854265 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.956287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.956342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.956355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.956377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:12 crc kubenswrapper[4809]: I1205 11:09:12.956391 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:12Z","lastTransitionTime":"2025-12-05T11:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.058724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.058786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.058800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.058817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.058829 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.161888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.161933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.161952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.161969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.161980 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.264558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.264613 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.264623 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.264651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.264660 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.367433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.367488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.367498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.367511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.367520 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.470749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.470847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.470869 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.470894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.470911 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.575127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.575204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.575226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.575255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.575285 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.678841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.678927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.678958 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.679062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.679157 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.782153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.782195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.782207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.782225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.782238 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.872133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:13 crc kubenswrapper[4809]: E1205 11:09:13.872306 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.872799 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:13 crc kubenswrapper[4809]: E1205 11:09:13.872986 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.873076 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.873158 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:13 crc kubenswrapper[4809]: E1205 11:09:13.873317 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:13 crc kubenswrapper[4809]: E1205 11:09:13.873446 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.888032 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.888529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.888552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.888575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.888590 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.992094 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.992184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.992197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.992218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:13 crc kubenswrapper[4809]: I1205 11:09:13.992253 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:13Z","lastTransitionTime":"2025-12-05T11:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.095874 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.095925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.095937 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.095954 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.095967 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.192796 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.198739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.198792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.198807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.198826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.198839 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.203616 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.209122 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.221646 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.236249 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.251486 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.275222 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.291393 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.303343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.303389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.303400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.303417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.303428 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.306506 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.318417 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.332234 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.346331 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.358896 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.374632 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.389538 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.405955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.406002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.406012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.406028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.406040 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.417007 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"message\\\":\\\"k controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:04.983664 6237 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983674 6237 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1205 11:09:04.983682 6237 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1205 11:09:04.983688 6237 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983689 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-n79nb\\\\nI1205 11:09:04.983687 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-s4hls\\\\nI1205 11:09:04.983697 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1205 11:09:04.983703 6237 obj_retry.go:365] Adding new object\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.430190 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.453592 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.472601 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.508759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.508828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.508853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.508883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.508905 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.611418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.611457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.611467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.611485 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.611496 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.713870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.713910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.713918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.713933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.713942 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.816495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.816520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.816528 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.816543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.816555 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.891611 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.912294 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.918759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.918809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.918820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.918840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.918913 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:14Z","lastTransitionTime":"2025-12-05T11:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.928411 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.939215 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.953880 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:14 crc kubenswrapper[4809]: I1205 11:09:14.994516 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:14Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.021953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.022060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.022077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.022260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.022270 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.022622 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.058585 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.069791 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.089182 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"message\\\":\\\"k controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:04.983664 6237 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983674 6237 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1205 11:09:04.983682 6237 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1205 11:09:04.983688 6237 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983689 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-n79nb\\\\nI1205 11:09:04.983687 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-s4hls\\\\nI1205 11:09:04.983697 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1205 11:09:04.983703 6237 obj_retry.go:365] Adding new object\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.103769 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.119487 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.125183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.125234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.125242 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.125257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.125266 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.134116 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.146730 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.163108 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.176666 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.192367 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.206673 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:15Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.228269 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.228311 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.228321 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.228336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.228351 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.330962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.331025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.331043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.331066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.331083 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.433586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.433617 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.433625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.433649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.433949 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.536527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.536561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.536571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.536586 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.536596 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.639589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.640759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.640811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.640835 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.640852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.720452 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.720614 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:09:47.720585184 +0000 UTC m=+83.111561782 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.743469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.743565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.743584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.743674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.743700 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.822349 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.822462 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.822523 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822554 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.822575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822698 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:47.822664799 +0000 UTC m=+83.213641407 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822774 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822794 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822835 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822862 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822889 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822867 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:47.822841634 +0000 UTC m=+83.213818232 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822928 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.822957 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.823004 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:47.822945017 +0000 UTC m=+83.213921575 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.823045 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:47.823021899 +0000 UTC m=+83.213998487 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.846506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.846550 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.846561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.846577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.846587 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.871444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.871467 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.871459 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.871444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.871601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.871779 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.871952 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:15 crc kubenswrapper[4809]: E1205 11:09:15.872032 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.949706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.949776 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.949786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.949801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:15 crc kubenswrapper[4809]: I1205 11:09:15.949814 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:15Z","lastTransitionTime":"2025-12-05T11:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.051693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.051741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.051756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.051774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.051785 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.155001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.155061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.155077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.155100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.155118 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.257353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.257401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.257416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.257438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.257454 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.360092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.360157 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.360171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.360231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.360255 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.463369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.463421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.463433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.463448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.463457 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.565973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.566015 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.566028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.566045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.566058 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.669105 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.669153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.669163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.669183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.669195 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.771691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.771731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.771740 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.771756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.771766 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.875611 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.875700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.875713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.875729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.875742 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.978527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.978593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.978605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.978627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:16 crc kubenswrapper[4809]: I1205 11:09:16.978691 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:16Z","lastTransitionTime":"2025-12-05T11:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.081130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.081184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.081203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.081226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.081244 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.184037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.184087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.184102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.184121 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.184136 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.287160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.287235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.287252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.287277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.287298 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.390246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.390286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.390297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.390313 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.390323 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.492279 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.492316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.492326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.492342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.492354 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.595855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.595921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.595940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.595964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.595984 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.699989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.700065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.700082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.700108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.700128 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.803277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.803338 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.803352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.803370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.803382 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.871475 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.871583 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.871650 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.871678 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:17 crc kubenswrapper[4809]: E1205 11:09:17.871724 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:17 crc kubenswrapper[4809]: E1205 11:09:17.871823 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:17 crc kubenswrapper[4809]: E1205 11:09:17.871975 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:17 crc kubenswrapper[4809]: E1205 11:09:17.872181 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.906659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.906711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.906725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.906744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:17 crc kubenswrapper[4809]: I1205 11:09:17.906764 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:17Z","lastTransitionTime":"2025-12-05T11:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.009743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.009803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.009815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.009833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.009844 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.113190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.113244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.113253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.113271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.113284 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.216955 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.217023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.217038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.217064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.217081 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.321091 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.321136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.321146 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.321167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.321187 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.424144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.424200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.424217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.424240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.424259 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.526622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.526677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.526686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.526700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.526712 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.617286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.617359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.617381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.617410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.617430 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: E1205 11:09:18.631187 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:18Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.637356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.637394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.637404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.637418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.637427 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: E1205 11:09:18.648778 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:18Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.652175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.652210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.652218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.652232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.652242 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: E1205 11:09:18.663081 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:18Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.666702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.666744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.666755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.666773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.666784 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: E1205 11:09:18.678577 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:18Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.681953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.681995 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.682007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.682025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.682036 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: E1205 11:09:18.693062 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:18Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:18 crc kubenswrapper[4809]: E1205 11:09:18.693219 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.694759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.694792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.694804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.694821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.694833 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.797065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.797350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.797358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.797372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.797381 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.900334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.900396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.900410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.900433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:18 crc kubenswrapper[4809]: I1205 11:09:18.900449 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:18Z","lastTransitionTime":"2025-12-05T11:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.003507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.003549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.003559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.003576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.003587 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.106129 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.106190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.106199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.106217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.106229 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.213917 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.213970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.213982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.214000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.214013 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.316513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.316575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.316590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.316616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.316675 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.420022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.420085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.420097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.420118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.420131 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.524136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.524196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.524213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.524230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.524240 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.627042 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.627108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.627120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.627144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.627157 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.729368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.729415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.729427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.729445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.729458 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.832116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.832165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.832181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.832203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.832219 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.866002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:19 crc kubenswrapper[4809]: E1205 11:09:19.866227 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:19 crc kubenswrapper[4809]: E1205 11:09:19.866341 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:09:35.866313424 +0000 UTC m=+71.257289982 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.871096 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:19 crc kubenswrapper[4809]: E1205 11:09:19.871251 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.871331 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:19 crc kubenswrapper[4809]: E1205 11:09:19.871397 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.871911 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.872074 4809 scope.go:117] "RemoveContainer" containerID="c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.871920 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:19 crc kubenswrapper[4809]: E1205 11:09:19.872161 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:19 crc kubenswrapper[4809]: E1205 11:09:19.872344 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.936007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.936062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.936145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.936177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:19 crc kubenswrapper[4809]: I1205 11:09:19.936199 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:19Z","lastTransitionTime":"2025-12-05T11:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.039133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.039159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.039169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.039184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.039195 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.142356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.142398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.142407 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.142422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.142431 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.244965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.245012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.245023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.245041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.245052 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.281826 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/1.log" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.283814 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.284181 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.297253 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.314303 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.326589 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.337074 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.347816 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.348108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.348152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.348164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.348180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.348191 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.374562 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"message\\\":\\\"k controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:04.983664 6237 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983674 6237 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1205 11:09:04.983682 6237 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1205 11:09:04.983688 6237 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983689 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-n79nb\\\\nI1205 11:09:04.983687 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-s4hls\\\\nI1205 11:09:04.983697 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1205 11:09:04.983703 6237 obj_retry.go:365] Adding new object\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.384618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.397435 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.413930 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.426963 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.441848 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.450345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.450376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.450384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.450397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.450406 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.453245 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.469518 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.481211 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.492804 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.505046 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.516728 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.528361 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.554186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.554236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.554248 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.554268 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.554282 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.656587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.656656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.656669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.656687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.656699 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.759467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.759502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.759511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.759524 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.759533 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.861957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.862013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.862024 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.862041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.862053 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.964830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.964892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.964909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.964929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:20 crc kubenswrapper[4809]: I1205 11:09:20.964945 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:20Z","lastTransitionTime":"2025-12-05T11:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.067703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.067774 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.067792 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.067815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.067832 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.171116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.171156 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.171165 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.171181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.171190 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.274502 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.274551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.274563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.274582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.274594 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.287978 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/2.log" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.288680 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/1.log" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.291225 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f" exitCode=1 Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.291272 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.291328 4809 scope.go:117] "RemoveContainer" containerID="c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.292029 4809 scope.go:117] "RemoveContainer" containerID="8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f" Dec 05 11:09:21 crc kubenswrapper[4809]: E1205 11:09:21.292216 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.305330 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.322348 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.338047 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.352392 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.369161 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.377255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.377301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.377313 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.377332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.377348 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.394207 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.410338 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.426170 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.436821 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.452053 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9b2fa248d6e010a9cf2fbc0743bb973da2e347b8d50cf5ee27ce59cb4c7cdf9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"message\\\":\\\"k controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:04Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:04.983664 6237 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983674 6237 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1205 11:09:04.983682 6237 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1205 11:09:04.983688 6237 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1205 11:09:04.983689 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-n79nb\\\\nI1205 11:09:04.983687 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-s4hls\\\\nI1205 11:09:04.983697 6237 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1205 11:09:04.983703 6237 obj_retry.go:365] Adding new object\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.462941 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.477915 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.480415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.480466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.480483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.480500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.480514 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.491673 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.503222 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.517356 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.529895 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.543805 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.557585 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:21Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.582713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.582786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.582799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.582816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.582828 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.685235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.685488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.685499 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.685515 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.685527 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.788801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.788866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.788878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.788897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.788909 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.871348 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.871398 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.871399 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.871457 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:21 crc kubenswrapper[4809]: E1205 11:09:21.871528 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:21 crc kubenswrapper[4809]: E1205 11:09:21.871755 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:21 crc kubenswrapper[4809]: E1205 11:09:21.871839 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:21 crc kubenswrapper[4809]: E1205 11:09:21.872012 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.891450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.891521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.891545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.891572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.891590 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.995607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.995664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.995673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.995686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:21 crc kubenswrapper[4809]: I1205 11:09:21.995696 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:21Z","lastTransitionTime":"2025-12-05T11:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.099128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.099197 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.099219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.099247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.099268 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.201926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.201999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.202012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.202030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.202042 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.303834 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/2.log" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.303929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.303974 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.303987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.304003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.304017 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.307835 4809 scope.go:117] "RemoveContainer" containerID="8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f" Dec 05 11:09:22 crc kubenswrapper[4809]: E1205 11:09:22.307967 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.325333 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.341956 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.357731 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.373863 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.387009 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.401273 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.407433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.407471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.407482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.407498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.407508 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.415573 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.431737 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.452856 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.465411 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.476172 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.487181 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.506723 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.509479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.509523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.509538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.509561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.509578 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.518664 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.532216 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.545127 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.565412 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.578987 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:22Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.612317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.612350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.612358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.612372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.612380 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.715425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.715490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.715543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.715560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.715571 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.818479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.818566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.818594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.818619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.818683 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.921455 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.921557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.921575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.921602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:22 crc kubenswrapper[4809]: I1205 11:09:22.921615 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:22Z","lastTransitionTime":"2025-12-05T11:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.024885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.024940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.024949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.024962 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.024971 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.127625 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.127712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.127724 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.127739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.127750 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.230797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.230843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.230854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.230869 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.230879 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.335107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.335225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.335238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.335255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.335264 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.438793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.438849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.438865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.438887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.438903 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.540946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.540986 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.541000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.541016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.541025 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.644459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.644604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.644670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.644779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.644807 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.748619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.748713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.748730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.748752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.748771 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.850577 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.850612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.850622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.850654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.850667 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.872180 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.872216 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.872221 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.872194 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:23 crc kubenswrapper[4809]: E1205 11:09:23.872295 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:23 crc kubenswrapper[4809]: E1205 11:09:23.872494 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:23 crc kubenswrapper[4809]: E1205 11:09:23.872592 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:23 crc kubenswrapper[4809]: E1205 11:09:23.872757 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.953406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.953446 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.953462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.953476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:23 crc kubenswrapper[4809]: I1205 11:09:23.953486 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:23Z","lastTransitionTime":"2025-12-05T11:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.056444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.056523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.056535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.056549 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.056560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.159668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.159704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.159716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.159731 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.159741 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.263596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.263697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.263721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.263750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.263772 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.366408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.366458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.366471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.366487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.366500 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.470551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.470849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.470861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.470877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.470888 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.573194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.573258 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.573281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.573310 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.573366 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.677060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.677092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.677101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.677113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.677123 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.780184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.780238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.780254 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.780280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.780298 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.883346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.883376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.883384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.883397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.883406 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.885156 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.897969 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.910534 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.922225 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.933467 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.944689 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.964340 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.976340 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.986000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.986038 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.986047 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.986064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.986075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:24Z","lastTransitionTime":"2025-12-05T11:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.987019 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:24 crc kubenswrapper[4809]: I1205 11:09:24.997441 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:24Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.016964 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.028886 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.041933 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.055773 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.069761 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.088390 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.088844 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.088888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.088900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.088919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.088933 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.102596 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.117547 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:25Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.191303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.191347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.191358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.191376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.191389 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.293394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.293452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.293461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.293474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.293503 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.395344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.395411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.395420 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.395434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.395443 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.505558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.505608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.505620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.505667 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.505691 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.608160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.608202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.608213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.608230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.608243 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.711793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.711839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.711848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.711862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.711872 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.814970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.815002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.815010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.815022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.815030 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.871370 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.871450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.871517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:25 crc kubenswrapper[4809]: E1205 11:09:25.871515 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.871559 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:25 crc kubenswrapper[4809]: E1205 11:09:25.871670 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:25 crc kubenswrapper[4809]: E1205 11:09:25.871748 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:25 crc kubenswrapper[4809]: E1205 11:09:25.871864 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.917425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.917457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.917466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.917478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:25 crc kubenswrapper[4809]: I1205 11:09:25.917487 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:25Z","lastTransitionTime":"2025-12-05T11:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.020074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.020102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.020114 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.020130 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.020142 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.122062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.122107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.122119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.122135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.122146 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.224752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.224799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.224808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.224822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.224833 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.327527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.327572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.327584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.327600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.327611 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.430007 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.430593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.430606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.430621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.430652 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.533451 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.533495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.533506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.533521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.533532 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.636389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.636483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.636494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.636521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.636540 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.739221 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.739271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.739283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.739301 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.739315 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.841472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.841545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.841557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.841598 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.841613 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.945061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.945112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.945127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.945147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:26 crc kubenswrapper[4809]: I1205 11:09:26.945159 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:26Z","lastTransitionTime":"2025-12-05T11:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.047405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.047462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.047474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.047492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.047503 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.150163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.150212 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.150223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.150238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.150250 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.252749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.252820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.252831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.252851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.252863 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.356145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.356211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.356226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.356251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.356267 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.458913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.458978 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.458990 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.459005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.459016 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.561676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.561934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.562040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.562122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.562196 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.668842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.668879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.668889 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.668902 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.668912 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.772662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.772701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.772714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.772730 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.772742 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.871771 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.871770 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.872288 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:27 crc kubenswrapper[4809]: E1205 11:09:27.872431 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.872310 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:27 crc kubenswrapper[4809]: E1205 11:09:27.872809 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:27 crc kubenswrapper[4809]: E1205 11:09:27.872837 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:27 crc kubenswrapper[4809]: E1205 11:09:27.872894 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.874718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.875025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.875291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.875604 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.875868 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.979379 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.979415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.979425 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.979438 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:27 crc kubenswrapper[4809]: I1205 11:09:27.979446 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:27Z","lastTransitionTime":"2025-12-05T11:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.082155 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.082188 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.082198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.082211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.082220 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.185340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.185384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.185397 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.185416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.185428 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.287932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.288266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.288277 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.288289 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.288298 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.390946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.390988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.390997 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.391028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.391039 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.493422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.493458 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.493467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.493480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.493490 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.595566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.595603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.595614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.595645 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.595658 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.697755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.697800 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.697816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.697831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.697842 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.799434 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.799461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.799470 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.799483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.799492 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.902065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.902136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.902158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.902191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:28 crc kubenswrapper[4809]: I1205 11:09:28.902217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:28Z","lastTransitionTime":"2025-12-05T11:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.000395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.000448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.000460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.000479 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.000493 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.012691 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:29Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.017357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.017399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.017412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.017430 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.017441 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.030458 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:29Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.034145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.034183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.034193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.034216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.034226 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.044572 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:29Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.047965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.047998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.048008 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.048022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.048031 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.059517 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:29Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.063136 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.063173 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.063182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.063196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.063217 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.074171 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:29Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.074365 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.076135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.076175 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.076187 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.076204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.076215 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.178681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.178727 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.178740 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.178782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.178794 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.281359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.281411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.281421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.281435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.281444 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.384080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.384117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.384127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.384143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.384153 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.486607 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.486675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.486685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.486702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.486714 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.589500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.589540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.589548 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.589563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.589573 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.691697 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.691745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.691754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.691770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.691779 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.794391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.794433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.794447 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.794462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.794473 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.871269 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.871376 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.871427 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.871433 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.871509 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.871552 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.871600 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:29 crc kubenswrapper[4809]: E1205 11:09:29.871678 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.896854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.896898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.896912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.896927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:29 crc kubenswrapper[4809]: I1205 11:09:29.896946 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:29Z","lastTransitionTime":"2025-12-05T11:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.000790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.000840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.000851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.000868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.000880 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.103259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.103305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.103316 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.103474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.103586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.206851 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.206933 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.206957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.206988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.207011 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.310011 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.310052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.310063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.310079 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.310093 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.412219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.412260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.412270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.412285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.412295 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.514741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.514781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.514796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.514813 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.514824 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.617681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.617725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.617737 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.617754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.617768 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.721329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.721388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.721400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.721417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.721431 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.824002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.824031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.824039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.824052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.824061 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.926322 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.926365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.926377 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.926393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:30 crc kubenswrapper[4809]: I1205 11:09:30.926407 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:30Z","lastTransitionTime":"2025-12-05T11:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.029672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.029763 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.029787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.029814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.029840 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.132765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.132836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.132857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.132888 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.132916 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.235752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.235811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.235830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.235853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.235904 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.338657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.338701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.338745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.338760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.338772 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.441093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.441141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.441150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.441163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.441172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.543567 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.543600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.543610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.543624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.543656 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.645576 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.645870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.646057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.646145 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.646236 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.749048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.749106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.749122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.749143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.749160 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.851373 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.851418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.851427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.851441 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.851450 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.871722 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.871850 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.871985 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:31 crc kubenswrapper[4809]: E1205 11:09:31.871983 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.872047 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:31 crc kubenswrapper[4809]: E1205 11:09:31.872185 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:31 crc kubenswrapper[4809]: E1205 11:09:31.872224 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:31 crc kubenswrapper[4809]: E1205 11:09:31.872279 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.954200 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.954236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.954246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.954262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:31 crc kubenswrapper[4809]: I1205 11:09:31.954273 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:31Z","lastTransitionTime":"2025-12-05T11:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.056733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.056829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.056854 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.056883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.056906 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.159312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.159371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.159386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.159409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.159426 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.262199 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.262260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.262270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.262292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.262309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.365076 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.365166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.365189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.365214 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.365229 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.467593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.467648 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.467659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.467678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.467690 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.569758 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.569799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.569829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.569848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.569859 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.672614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.672737 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.672761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.672793 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.672830 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.775850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.775926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.775952 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.775984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.776006 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.878442 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.878505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.878521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.878546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.878563 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.980879 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.980919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.980930 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.980947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:32 crc kubenswrapper[4809]: I1205 11:09:32.980959 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:32Z","lastTransitionTime":"2025-12-05T11:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.083833 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.083870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.083881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.083896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.083909 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.186920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.186963 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.186975 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.186992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.187005 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.289751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.289787 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.289798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.289815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.289827 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.392721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.392892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.392918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.392940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.392961 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.495677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.495930 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.495949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.495972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.495989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.598821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.598885 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.598900 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.598923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.598939 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.701770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.701819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.701827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.701843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.701852 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.805746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.805799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.805811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.805829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.805844 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.872028 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:33 crc kubenswrapper[4809]: E1205 11:09:33.872190 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.872455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:33 crc kubenswrapper[4809]: E1205 11:09:33.872573 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.872794 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.872912 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:33 crc kubenswrapper[4809]: E1205 11:09:33.872958 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:33 crc kubenswrapper[4809]: E1205 11:09:33.873176 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.909001 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.909066 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.909086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.909113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:33 crc kubenswrapper[4809]: I1205 11:09:33.909131 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:33Z","lastTransitionTime":"2025-12-05T11:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.011769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.011818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.011826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.011843 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.011853 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.114402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.114448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.114463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.114483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.114498 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.217016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.217078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.217090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.217108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.217118 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.319820 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.319868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.319878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.319896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.319909 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.424453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.424488 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.424498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.424510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.424522 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.526513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.526557 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.526566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.526578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.526586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.629108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.629189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.629223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.629273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.629298 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.731767 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.731799 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.731808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.731823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.731834 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.835115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.835166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.835178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.835196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.835207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.884956 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.898998 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.912668 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.928127 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.938125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.938160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.938169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.938185 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.938196 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:34Z","lastTransitionTime":"2025-12-05T11:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.945124 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.956818 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.971333 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.983526 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:34 crc kubenswrapper[4809]: I1205 11:09:34.997999 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:34Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.009869 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.018509 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.036533 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.040460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.040506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.040519 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.040540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.040553 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.047315 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.059079 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.079672 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.093517 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.103998 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.112939 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:35Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.142670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.142702 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.142711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.142743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.142751 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.245283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.245329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.245339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.245356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.245366 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.348805 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.348859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.348877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.348898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.348915 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.451345 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.451392 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.451405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.451444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.451459 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.554680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.554745 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.554761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.554784 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.554802 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.657003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.657044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.657057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.657073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.657085 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.759597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.759649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.759659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.759672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.759680 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.862622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.862676 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.862684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.862699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.862708 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.872164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.872210 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:35 crc kubenswrapper[4809]: E1205 11:09:35.872363 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.872387 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.872426 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:35 crc kubenswrapper[4809]: E1205 11:09:35.872549 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:35 crc kubenswrapper[4809]: E1205 11:09:35.872619 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:35 crc kubenswrapper[4809]: E1205 11:09:35.872708 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.935500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:35 crc kubenswrapper[4809]: E1205 11:09:35.935674 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:35 crc kubenswrapper[4809]: E1205 11:09:35.935764 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:10:07.935745703 +0000 UTC m=+103.326722261 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.964584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.964652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.964663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.964681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:35 crc kubenswrapper[4809]: I1205 11:09:35.964695 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:35Z","lastTransitionTime":"2025-12-05T11:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.066940 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.066988 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.066998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.067017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.067031 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.169394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.169433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.169443 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.169461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.169474 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.271663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.271694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.271704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.271723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.271734 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.373830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.373877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.373894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.373912 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.373924 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.477088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.477162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.477184 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.477211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.477234 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.580348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.580399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.580414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.580431 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.580441 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.683408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.683459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.683471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.683485 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.683495 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.786261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.786305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.786318 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.786333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.786347 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.873044 4809 scope.go:117] "RemoveContainer" containerID="8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f" Dec 05 11:09:36 crc kubenswrapper[4809]: E1205 11:09:36.873232 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.887275 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.888426 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.888454 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.888466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.888483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.888492 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.992140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.992191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.992203 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.992219 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:36 crc kubenswrapper[4809]: I1205 11:09:36.992232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:36Z","lastTransitionTime":"2025-12-05T11:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.095103 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.095148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.095160 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.095178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.095193 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.197043 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.197081 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.197092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.197109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.197120 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.300074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.300116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.300131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.300148 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.300159 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.402462 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.402510 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.402521 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.402538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.402551 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.505202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.505248 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.505257 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.505273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.505283 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.608161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.608192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.608202 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.608217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.608228 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.710749 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.710789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.710803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.710826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.710837 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.813125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.813159 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.813168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.813180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.813188 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.871667 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.871690 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.871699 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.871704 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:37 crc kubenswrapper[4809]: E1205 11:09:37.871797 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:37 crc kubenswrapper[4809]: E1205 11:09:37.871894 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:37 crc kubenswrapper[4809]: E1205 11:09:37.872042 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:37 crc kubenswrapper[4809]: E1205 11:09:37.872154 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.916235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.916280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.916292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.916306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:37 crc kubenswrapper[4809]: I1205 11:09:37.916315 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:37Z","lastTransitionTime":"2025-12-05T11:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.019096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.019153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.019174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.019209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.019232 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.121943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.121982 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.122021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.122037 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.122049 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.227228 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.227261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.227270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.227287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.227297 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.330680 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.330743 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.330760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.330782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.330803 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.433560 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.433618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.433664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.433687 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.433704 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.537215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.537587 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.537779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.537944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.538094 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.642579 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.642612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.642622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.642655 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.642665 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.745227 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.745692 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.745842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.745983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.746097 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.848486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.848539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.848551 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.848568 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.848582 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.950786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.950829 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.950839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.950855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:38 crc kubenswrapper[4809]: I1205 11:09:38.950865 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:38Z","lastTransitionTime":"2025-12-05T11:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.053659 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.053713 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.053750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.053769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.053784 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.156921 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.156967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.156998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.157014 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.157026 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.259882 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.259918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.259927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.259939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.259948 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.335480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.335536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.335546 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.335563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.335575 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.359184 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/0.log" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.359266 4809 generic.go:334] "Generic (PLEG): container finished" podID="238d6f1b-f3b4-4b33-a33c-10a0c99e8178" containerID="ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da" exitCode=1 Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.359295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerDied","Data":"ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.359625 4809 scope.go:117] "RemoveContainer" containerID="ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.362815 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.367304 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.367341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.367352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.367369 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.367382 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.378642 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.382691 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.388976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.391617 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.391654 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.391670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.391682 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.395189 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.403489 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.405586 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.406931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.406989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.407005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.407023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.407033 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.419795 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.421358 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.426994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.427029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.427040 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.427056 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.427068 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.432229 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.438123 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.438292 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.440088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.440141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.440150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.440162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.440172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.447079 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:38Z\\\",\\\"message\\\":\\\"2025-12-05T11:08:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5\\\\n2025-12-05T11:08:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5 to /host/opt/cni/bin/\\\\n2025-12-05T11:08:53Z [verbose] multus-daemon started\\\\n2025-12-05T11:08:53Z [verbose] Readiness Indicator file check\\\\n2025-12-05T11:09:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.459598 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.472032 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"850c8814-2b0a-420c-a30a-0d663ad3d191\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6a8e3f6b18707b772ca7e0a55ec37e40c7ab09c75c6585b9715187f221505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.483781 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.496891 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.507621 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.518687 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.530523 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.542861 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.542910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.542923 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.542941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.542956 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.550006 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.561732 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.573152 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.584039 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.607105 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.622555 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:39Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.645036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.645078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.645090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.645107 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.645130 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.747486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.747589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.747603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.747620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.747643 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.850714 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.850753 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.850764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.850783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.850796 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.871539 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.871568 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.871625 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.871550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.871693 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.871784 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.871843 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:39 crc kubenswrapper[4809]: E1205 11:09:39.871906 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.952581 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.952620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.952657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.952674 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:39 crc kubenswrapper[4809]: I1205 11:09:39.952684 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:39Z","lastTransitionTime":"2025-12-05T11:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.055649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.056017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.056028 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.056046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.056057 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.158790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.158869 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.158895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.158927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.158950 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.261985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.262045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.262064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.262089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.262106 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.365020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.365073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.365096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.365127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.365145 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.367242 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/0.log" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.367325 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerStarted","Data":"65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.385584 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.402786 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:38Z\\\",\\\"message\\\":\\\"2025-12-05T11:08:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5\\\\n2025-12-05T11:08:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5 to /host/opt/cni/bin/\\\\n2025-12-05T11:08:53Z [verbose] multus-daemon started\\\\n2025-12-05T11:08:53Z [verbose] Readiness Indicator file check\\\\n2025-12-05T11:09:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.424262 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.446715 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.465395 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.468127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.468164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.468174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.468189 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.468198 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.486783 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.504569 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.517880 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"850c8814-2b0a-420c-a30a-0d663ad3d191\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6a8e3f6b18707b772ca7e0a55ec37e40c7ab09c75c6585b9715187f221505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.532271 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.550472 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.564964 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.573195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.573241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.573253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.573275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.573288 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.580000 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.591457 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.609439 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.620846 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.634100 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.655834 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.670168 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.675411 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.675732 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.675837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.675927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.676007 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.683884 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:40Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.778337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.778657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.778766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.778868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.778953 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.880806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.880839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.880849 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.880865 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.880895 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.983460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.983500 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.983508 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.983523 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:40 crc kubenswrapper[4809]: I1205 11:09:40.983536 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:40Z","lastTransitionTime":"2025-12-05T11:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.086131 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.086181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.086194 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.086211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.086224 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.188699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.188760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.188772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.188786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.188798 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.291051 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.291115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.291134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.291163 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.291182 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.393961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.394031 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.394053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.394083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.394108 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.495916 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.495979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.495993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.496012 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.496025 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.598490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.598566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.598585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.598602 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.598613 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.700944 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.700977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.700985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.700999 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.701009 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.803478 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.803527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.803539 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.803558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.803571 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.871815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:41 crc kubenswrapper[4809]: E1205 11:09:41.871951 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.871976 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.872075 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.871986 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:41 crc kubenswrapper[4809]: E1205 11:09:41.872129 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:41 crc kubenswrapper[4809]: E1205 11:09:41.872300 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:41 crc kubenswrapper[4809]: E1205 11:09:41.872444 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.905512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.905556 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.905569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.905585 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:41 crc kubenswrapper[4809]: I1205 11:09:41.905596 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:41Z","lastTransitionTime":"2025-12-05T11:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.008178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.008401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.008414 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.008432 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.008443 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.111412 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.111466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.111476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.111492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.111503 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.214883 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.214935 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.214949 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.214966 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.214979 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.317657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.317693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.317703 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.317718 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.317728 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.420118 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.420166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.420179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.420195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.420208 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.524089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.524172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.524196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.524225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.524249 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.626895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.626924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.626934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.626946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.626955 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.729867 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.729956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.729987 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.730016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.730038 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.832167 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.832232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.832252 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.832291 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.832326 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.934555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.934584 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.934593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.934605 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:42 crc kubenswrapper[4809]: I1205 11:09:42.934613 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:42Z","lastTransitionTime":"2025-12-05T11:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.036765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.036811 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.036822 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.036836 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.036845 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.139956 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.139989 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.140000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.140016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.140028 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.242755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.242790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.242801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.242816 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.242826 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.344837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.344887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.344897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.344915 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.344927 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.447647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.447690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.447701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.447719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.447733 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.550898 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.550947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.550957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.550974 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.550985 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.653226 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.653270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.653280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.653298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.653309 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.756337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.756396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.756415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.756439 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.756456 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.859174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.859232 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.859256 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.859285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.859306 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.871795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.871867 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.871886 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.871803 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:43 crc kubenswrapper[4809]: E1205 11:09:43.871999 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:43 crc kubenswrapper[4809]: E1205 11:09:43.872101 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:43 crc kubenswrapper[4809]: E1205 11:09:43.872189 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:43 crc kubenswrapper[4809]: E1205 11:09:43.872247 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.961868 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.961918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.961929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.961948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:43 crc kubenswrapper[4809]: I1205 11:09:43.961959 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:43Z","lastTransitionTime":"2025-12-05T11:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.064209 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.064259 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.064271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.064288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.064300 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.167565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.167669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.167694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.167721 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.167743 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.270501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.270575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.270609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.270677 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.270704 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.373085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.373127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.373135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.373150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.373159 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.475596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.475715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.475738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.475775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.475798 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.578600 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.578658 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.578671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.578690 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.578703 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.685147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.685205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.685223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.685247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.685264 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.788151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.788220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.788244 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.788275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.788297 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.891299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.891337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.891347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.891361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.891372 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.896613 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:44Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.912291 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:44Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.928576 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:44Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.957537 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:44Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.986544 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:44Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.994233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.994463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.994474 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.994491 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:44 crc kubenswrapper[4809]: I1205 11:09:44.994502 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:44Z","lastTransitionTime":"2025-12-05T11:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.008777 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.042874 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.057290 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.074741 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.087417 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.097470 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.097555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.097571 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.097595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.097610 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.101897 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:38Z\\\",\\\"message\\\":\\\"2025-12-05T11:08:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5\\\\n2025-12-05T11:08:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5 to /host/opt/cni/bin/\\\\n2025-12-05T11:08:53Z [verbose] multus-daemon started\\\\n2025-12-05T11:08:53Z [verbose] Readiness Indicator file check\\\\n2025-12-05T11:09:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.119557 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.133406 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.152329 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.163540 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"850c8814-2b0a-420c-a30a-0d663ad3d191\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6a8e3f6b18707b772ca7e0a55ec37e40c7ab09c75c6585b9715187f221505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.176871 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.187300 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.200924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.200992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.201006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.201022 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.201033 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.201563 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.220791 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:45Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.303764 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.303808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.303819 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.303862 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.303882 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.407071 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.407142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.407156 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.407181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.407196 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.510182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.510258 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.510271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.510296 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.510311 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.613112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.613193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.613207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.613231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.613245 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.716281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.716352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.716375 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.716404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.716488 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.819180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.819297 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.819319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.819344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.819361 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.871152 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.871319 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.871468 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:45 crc kubenswrapper[4809]: E1205 11:09:45.871503 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.871546 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:45 crc kubenswrapper[4809]: E1205 11:09:45.871719 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:45 crc kubenswrapper[4809]: E1205 11:09:45.871960 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:45 crc kubenswrapper[4809]: E1205 11:09:45.872121 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.922294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.922339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.922353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.922368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:45 crc kubenswrapper[4809]: I1205 11:09:45.922382 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:45Z","lastTransitionTime":"2025-12-05T11:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.025606 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.025722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.025744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.025770 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.025788 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.129483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.129547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.129566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.129590 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.129607 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.232261 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.232323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.232341 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.232368 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.232393 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.335913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.335992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.336009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.336034 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.336052 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.438558 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.438599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.438610 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.438651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.438665 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.541409 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.541460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.541471 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.541489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.541501 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.644067 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.644122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.644138 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.644161 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.644186 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.747675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.747751 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.747776 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.747804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.747821 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.850476 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.850573 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.850594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.850668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.850708 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.953402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.953440 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.953450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.953465 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:46 crc kubenswrapper[4809]: I1205 11:09:46.953476 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:46Z","lastTransitionTime":"2025-12-05T11:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.056211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.056263 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.056274 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.056290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.056301 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.158112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.158150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.158158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.158174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.158183 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.261575 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.261624 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.261649 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.261665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.261676 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.364324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.364428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.364460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.364492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.364514 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.467487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.467537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.467555 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.467574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.467587 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.570147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.570178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.570191 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.570207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.570220 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.672385 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.672445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.672463 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.672487 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.672505 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.756613 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.756826 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:51.756794666 +0000 UTC m=+147.147771234 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.775294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.775365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.775382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.775405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.775422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.857751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.857814 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.857843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.857868 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.857902 4809 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.857965 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.857970 4809 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.857979 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.857996 4809 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.857978 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:10:51.857961134 +0000 UTC m=+147.248937702 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.858036 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 11:10:51.858026816 +0000 UTC m=+147.249003374 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.858048 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 11:10:51.858041936 +0000 UTC m=+147.249018494 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.858138 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.858183 4809 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.858195 4809 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.858252 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 11:10:51.858235962 +0000 UTC m=+147.249212520 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.871708 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.871732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.871801 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.871708 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.871905 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.871969 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.872060 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:47 crc kubenswrapper[4809]: E1205 11:09:47.872169 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.881875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.881969 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.881985 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.882010 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.882032 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.985932 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.985998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.986020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.986050 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:47 crc kubenswrapper[4809]: I1205 11:09:47.986072 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:47Z","lastTransitionTime":"2025-12-05T11:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.089054 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.089100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.089112 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.089132 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.089148 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.191975 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.192027 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.192039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.192057 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.192070 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.295432 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.295492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.295538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.295562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.295580 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.398592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.398666 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.398679 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.398694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.398706 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.502072 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.502123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.502137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.502158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.502172 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.604789 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.604825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.604837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.604855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.604866 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.708669 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.708810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.708834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.708863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.708884 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.811208 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.811233 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.811241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.811253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.811264 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.913785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.914123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.914211 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.914305 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:48 crc kubenswrapper[4809]: I1205 11:09:48.914404 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:48Z","lastTransitionTime":"2025-12-05T11:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.017151 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.017489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.017582 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.017712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.017782 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.121080 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.121147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.121170 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.121201 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.121224 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.223950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.224006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.224017 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.224036 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.224055 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.327218 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.327290 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.327317 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.327347 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.327371 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.430133 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.430166 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.430179 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.430195 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.430207 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.532972 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.533030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.533045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.533102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.533117 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.635853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.635892 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.635909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.635926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.635938 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.713924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.713994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.714018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.714048 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.714075 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.734328 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.739918 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.739957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.739968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.739984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.739997 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.752839 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.756143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.756174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.756182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.756198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.756208 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.769396 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.773875 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.773927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.773943 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.773968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.773986 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.789654 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.793691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.793734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.793750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.793771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.793788 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.808597 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:49Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.808859 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.810700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.810740 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.810755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.810773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.810789 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.871876 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.871948 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.872048 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.872043 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.872101 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.872199 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.872297 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:49 crc kubenswrapper[4809]: E1205 11:09:49.872376 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.873326 4809 scope.go:117] "RemoveContainer" containerID="8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.915275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.915786 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.916098 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.916300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:49 crc kubenswrapper[4809]: I1205 11:09:49.916482 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:49Z","lastTransitionTime":"2025-12-05T11:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.020681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.020845 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.020980 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.021093 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.021174 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.123941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.124005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.124030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.124064 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.124087 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.227115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.227399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.227448 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.227466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.227476 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.331055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.331110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.331127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.331150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.331166 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.404128 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/2.log" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.407181 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.407682 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.421792 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.431078 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"850c8814-2b0a-420c-a30a-0d663ad3d191\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6a8e3f6b18707b772ca7e0a55ec37e40c7ab09c75c6585b9715187f221505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.433643 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.433678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.433686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.433701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.433710 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.441277 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.450391 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.461132 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.479401 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.498918 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.519006 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.529988 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.535193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.535220 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.535229 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.535245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.535257 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.546849 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.556282 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.565697 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.583303 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.594752 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.609317 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.624203 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.637830 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.637886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.637895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.637909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.637918 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.638706 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:38Z\\\",\\\"message\\\":\\\"2025-12-05T11:08:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5\\\\n2025-12-05T11:08:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5 to /host/opt/cni/bin/\\\\n2025-12-05T11:08:53Z [verbose] multus-daemon started\\\\n2025-12-05T11:08:53Z [verbose] Readiness Indicator file check\\\\n2025-12-05T11:09:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.650543 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.662851 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:50Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.739886 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.739946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.739959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.739977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.739989 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.843055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.843109 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.843124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.843147 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.843183 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.945469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.945514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.945522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.945535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:50 crc kubenswrapper[4809]: I1205 11:09:50.945543 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:50Z","lastTransitionTime":"2025-12-05T11:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.048003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.048049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.048061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.048077 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.048088 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.150418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.150675 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.150688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.150707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.150718 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.252707 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.252741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.252759 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.252775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.252787 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.356580 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.356672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.356686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.356705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.356720 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.458619 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.458673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.458685 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.458700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.458712 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.561847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.562102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.562180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.562223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.562249 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.666292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.666350 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.666361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.666381 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.666396 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.769283 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.769362 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.769405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.769435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.769450 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.871247 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.871283 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.871283 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.871788 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:51 crc kubenswrapper[4809]: E1205 11:09:51.871535 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:51 crc kubenswrapper[4809]: E1205 11:09:51.872028 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:51 crc kubenswrapper[4809]: E1205 11:09:51.872124 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:51 crc kubenswrapper[4809]: E1205 11:09:51.872196 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.873593 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.873698 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.873715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.873733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.873745 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.976111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.976186 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.976205 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.976235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:51 crc kubenswrapper[4809]: I1205 11:09:51.976254 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:51Z","lastTransitionTime":"2025-12-05T11:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.079314 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.079468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.079490 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.079514 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.079531 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.183164 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.183206 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.183215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.183231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.183242 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.286280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.286349 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.286363 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.286389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.286406 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.390039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.390089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.390100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.390116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.390128 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.414831 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/3.log" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.415572 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/2.log" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.419127 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" exitCode=1 Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.419205 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.419264 4809 scope.go:117] "RemoveContainer" containerID="8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.420759 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:09:52 crc kubenswrapper[4809]: E1205 11:09:52.421109 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.444348 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.458567 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.474167 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.485092 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.492106 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.492180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.492198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.492223 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.492239 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.497393 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.521654 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.535066 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.549317 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.561112 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.581870 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:51Z\\\",\\\"message\\\":\\\" for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 11:09:50.753564 6853 services_controller.go:452] Built service openshift-controller-manager-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1205 11:09:50.753563 6853 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:50.753575 6853 services_controller.go:453] Built service openshift-controller-manager-operator/metrics template LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.596100 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.601495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.601543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.601554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.601570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.601580 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.614854 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.631175 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.643226 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.659400 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.671607 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.684311 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:38Z\\\",\\\"message\\\":\\\"2025-12-05T11:08:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5\\\\n2025-12-05T11:08:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5 to /host/opt/cni/bin/\\\\n2025-12-05T11:08:53Z [verbose] multus-daemon started\\\\n2025-12-05T11:08:53Z [verbose] Readiness Indicator file check\\\\n2025-12-05T11:09:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.697789 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.704300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.704373 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.704398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.704421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.704436 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.709296 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"850c8814-2b0a-420c-a30a-0d663ad3d191\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6a8e3f6b18707b772ca7e0a55ec37e40c7ab09c75c6585b9715187f221505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:52Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.807029 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.807082 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.807102 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.807128 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.807146 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.909620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.909725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.909740 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.909761 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:52 crc kubenswrapper[4809]: I1205 11:09:52.909777 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:52Z","lastTransitionTime":"2025-12-05T11:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.013234 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.013337 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.013359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.013390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.013413 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.116210 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.116255 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.116265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.116278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.116287 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.219062 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.219100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.219111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.219127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.219139 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.321660 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.321704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.321715 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.321734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.321746 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.423814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.423946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.424061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.424095 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.424165 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.425827 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/3.log" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.527309 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.527386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.527404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.527433 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.527454 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.629807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.629870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.629890 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.629920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.629940 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.733023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.733105 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.733124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.733150 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.733170 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.836941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.837023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.837053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.837087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.837106 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.871857 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.871895 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.872001 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.872344 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:53 crc kubenswrapper[4809]: E1205 11:09:53.872507 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:53 crc kubenswrapper[4809]: E1205 11:09:53.872899 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:53 crc kubenswrapper[4809]: E1205 11:09:53.873068 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:53 crc kubenswrapper[4809]: E1205 11:09:53.873169 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.939673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.939737 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.939754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.939776 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:53 crc kubenswrapper[4809]: I1205 11:09:53.939794 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:53Z","lastTransitionTime":"2025-12-05T11:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.043115 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.043513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.043712 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.043920 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.044088 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.147870 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.147974 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.147994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.148016 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.148033 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.250124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.250460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.250609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.250803 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.250994 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.354427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.354845 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.355061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.355278 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.355483 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.458336 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.458391 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.458404 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.458423 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.458437 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.560895 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.560948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.560970 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.560984 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.560995 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.663896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.663948 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.663968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.663996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.664018 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.766403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.766485 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.766516 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.766562 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.766599 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.869569 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.869627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.869668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.869689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.869704 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.886977 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87517ccc-4f19-430e-9d03-d71da4cbe196\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5ad8425077f5c0c93271501831454cdac0810a858201e4cf4e3259fcc447299\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbtl8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s4hls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.904119 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n79nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238d6f1b-f3b4-4b33-a33c-10a0c99e8178\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:38Z\\\",\\\"message\\\":\\\"2025-12-05T11:08:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5\\\\n2025-12-05T11:08:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00dd31ef-b8ab-4b23-969b-bbcd42609ad5 to /host/opt/cni/bin/\\\\n2025-12-05T11:08:53Z [verbose] multus-daemon started\\\\n2025-12-05T11:08:53Z [verbose] Readiness Indicator file check\\\\n2025-12-05T11:09:38Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jgjfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n79nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.919874 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e8da43f-5236-4877-9010-364b45224889\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.935094 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7ab2f783d56e44be78ce20ade42aa51d5aa17774835b05f1c9891dfec20d05d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.948438 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e33349347c025523a3bf7201c9c1e4ee8bd454a374de4360a73c47335d75409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.964618 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38775c19-63c3-49a0-974e-5629199dcd55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39cf558ffcde744a30bacf3962fea0a79719db18bd875aa0a98f1f3f23b98692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5f60c8317ed455e4220370375476b2f5ff1f0c704a43ce05e7d62fc1fb49505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc87a85f53c238120f7be194cacdb0eeb61e8c3d03a488f79689a05a78eb7b1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf8279c28c99dd57e6c870a938d0f0726863dca806a9c28d44c6d5479f332707\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b25422bec0fea780de9636fe072e765f4c7a82c6e0833e8e5121ede9a505815e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd12813aeb41746f73bc66b6223e67995ebf222c2b5cfc56b6f3595fea976ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8aa96f15fc3d4d36236ddc4aea9fd17e6330817d184b4da6e7094522d8b832a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dwrpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwhnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.972722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.972757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.972765 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.972779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.972788 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:54Z","lastTransitionTime":"2025-12-05T11:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.977901 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1cbe6e95-f6ee-44a9-89aa-efbbaa3a3c06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e720f3722cc8cb4bc7b9acb011d25c55bc0f4ab18ea878ab4903d5380569c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42370ca404a7a2c27b611f96c23b5a1c4323056880815a7fe9c8763447ea4505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76acdd79ae49c22d8cc2929331b423c21a23b69e494aec5abbc85d796f866b73\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:54 crc kubenswrapper[4809]: I1205 11:09:54.991192 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"850c8814-2b0a-420c-a30a-0d663ad3d191\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a6a8e3f6b18707b772ca7e0a55ec37e40c7ab09c75c6585b9715187f221505a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b8b94391ea1430285e440dc81769b9092805d43f1962a88d51358c7bb93b52f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:54Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.007369 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.020324 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a915a4db4f20bee24725c0a0cc0e90c87a3b5cd59cfe3bf94b3195683030d62a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://805c923515e951e79cab515ee2ccd66cf7826841d2c17ae8d1cda9e95ec01a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.032817 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66d25413-5343-45fe-acb6-3d72fb5a4e3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0ac927ec7411b3843b8e1fd9990cfdccb6dc684addea7e9b2d3d645c41ee94a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c27166745cf583c573a5dc5d235a3fad21254a74947065a7e6ba70ad0fd39f0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8h9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ssnzv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.043681 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-skrvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:09:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxkn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.052786 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rjsd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7d9fdbc-ab41-4867-8e71-807ff609fb1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845d1864839adee61f810c2411bd2a7f9581cf21d59b1c26ee19c319914510a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pfqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rjsd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.071033 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dbc2d525279061a459f8c74a22e4e09bf5d3770a64d71552c28e627db795d3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:21Z\\\",\\\"message\\\":\\\"d == {fe46cb89-4e54-4175-a112-1c5224cd299e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:20.631407 6444 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1205 11:09:20.631408 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:20Z is after 2025-08-24T17:21:41Z]\\\\nI1205 11:09:20.631412 6444 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1205 11:09:20.631417 6444 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T11:09:51Z\\\",\\\"message\\\":\\\" for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 11:09:50.753564 6853 services_controller.go:452] Built service openshift-controller-manager-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1205 11:09:50.753563 6853 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 11:09:50.753575 6853 services_controller.go:453] Built service openshift-controller-manager-operator/metrics template LB for\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T11:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4kb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cdvk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.075088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.075127 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.075135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.075153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.075163 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.081419 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-l5kgl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb95f25-779a-4461-82d1-b7ea053eb2c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c0b84f390abec32e62a0ad3be3673ba21d17183a2cf28895328ce11c45ff6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhmm6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-l5kgl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.093896 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de4c0896-db8d-485a-a3a4-050fa007ac68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cbbbdc02603e6bafa11039ef617b2faa73372ca0c29f5fa64b6939be79f2da6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9a2ba1df100997089fdc69013cf722b6e53e1a41ebdbdc7d42f75cab7752868\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2449669378dac6b67c3e743ae198dcc1b352f933209f93e80ad8855bd085cf5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01e6fb0b40a4b055a401056cda42fd5c8a038da578bd9cee670e7b185984008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.112841 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16116f5a-bffd-4b0a-9bd5-da8647d9b59d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a1cd02641440e26e9a92dc1ea8be16b4dd2a861c4056b62c7f946a9996163b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6305de03a1081141f7fc3b9fb444d1ecae94d8a28ee2e32e9015824c72a87e53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d06f8d598973906726b79020aa078dd6e75647512d550eef105eb3f5b543ed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7e4627a0bf930573f59f85fb85dd849805489cbd94ae45a7944d8f21d9a1a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86b38af3431bca082339cdb625ec60a46f8c1af62f9034cdc0d7a6eb498a02a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T11:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d16ef00d43b1729f789ec994417083c47ab90935d8ec8c34e780f18b3201e373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://303ea1199add64e979a6b4108e4a98001b7ce051bd327facaf8d87f565097a9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099dc0b782721b2effd94d32c49c41a23f0737b33f3fc54bd397ce0e9ea5d89b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T11:08:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T11:08:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T11:08:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.124544 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.136908 4809 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T11:08:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:55Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.179181 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.179225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.179237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.179253 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.179267 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.281656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.281686 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.281694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.281709 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.281720 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.383960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.384020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.384045 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.384074 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.384097 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.487396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.487428 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.487450 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.487466 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.487477 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.590090 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.590143 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.590158 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.590177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.590546 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.692701 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.692771 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.692796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.692824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.692847 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.796282 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.796346 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.796370 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.796399 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.796422 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.871160 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.871200 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.871219 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.871280 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:55 crc kubenswrapper[4809]: E1205 11:09:55.871278 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:55 crc kubenswrapper[4809]: E1205 11:09:55.871385 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:55 crc kubenswrapper[4809]: E1205 11:09:55.871469 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:55 crc kubenswrapper[4809]: E1205 11:09:55.871521 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.898671 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.898711 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.898723 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.898769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:55 crc kubenswrapper[4809]: I1205 11:09:55.898781 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:55Z","lastTransitionTime":"2025-12-05T11:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.002241 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.002333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.002358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.002394 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.002419 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.105460 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.105527 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.105540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.105559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.105575 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.209356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.209408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.209429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.209452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.209469 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.312672 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.312734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.312755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.312780 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.312799 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.415725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.415767 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.415781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.415798 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.415811 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.518566 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.518684 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.518708 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.518739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.518764 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.622213 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.622264 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.622276 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.622294 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.622307 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.726049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.726140 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.726162 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.726192 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.726214 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.829695 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.829779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.829794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.829815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.829829 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.933906 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.933977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.934030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.934055 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:56 crc kubenswrapper[4809]: I1205 11:09:56.934071 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:56Z","lastTransitionTime":"2025-12-05T11:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.036947 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.036993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.037003 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.037018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.037027 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.139704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.139785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.139817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.139850 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.139876 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.242960 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.242998 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.243009 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.243025 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.243039 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.346177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.346240 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.346250 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.346270 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.346287 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.449418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.449461 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.449472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.449492 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.449505 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.552013 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.552058 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.552069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.552085 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.552096 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.654622 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.654682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.654694 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.654710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.654719 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.756752 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.756808 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.756826 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.756847 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.756864 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.859993 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.860041 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.860052 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.860069 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.860085 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.871290 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.871326 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.871372 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.871598 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:57 crc kubenswrapper[4809]: E1205 11:09:57.871586 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:57 crc kubenswrapper[4809]: E1205 11:09:57.871717 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:57 crc kubenswrapper[4809]: E1205 11:09:57.871977 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:57 crc kubenswrapper[4809]: E1205 11:09:57.872182 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.962802 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.962837 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.962846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.962859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:57 crc kubenswrapper[4809]: I1205 11:09:57.962868 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:57Z","lastTransitionTime":"2025-12-05T11:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.065700 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.065766 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.065781 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.065807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.065822 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.169343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.169422 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.169437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.169467 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.169483 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.272750 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.272804 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.272821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.272842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.272858 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.375863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.375899 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.375910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.375925 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.375938 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.478018 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.478063 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.478086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.478100 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.478110 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.580957 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.580994 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.581005 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.581023 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.581035 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.683815 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.683881 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.683904 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.683931 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.683953 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.786512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.786561 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.786574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.786592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.786604 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.888323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.888419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.888445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.888494 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.888526 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.991444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.991498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.991512 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.991534 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:58 crc kubenswrapper[4809]: I1205 11:09:58.991545 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:58Z","lastTransitionTime":"2025-12-05T11:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.095235 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.095760 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.095775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.095794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.095808 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.199119 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.199180 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.199198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.199225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.199242 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.301852 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.301897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.301909 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.301924 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.301935 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.404821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.404897 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.404913 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.404941 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.404959 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.507739 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.507783 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.507794 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.507809 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.507820 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.610366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.610405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.610417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.610429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.610438 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.713405 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.713821 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.713834 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.713857 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.713872 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.817262 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.817319 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.817333 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.817353 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.817365 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.872182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.872218 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.872319 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:09:59 crc kubenswrapper[4809]: E1205 11:09:59.872344 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.872406 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:09:59 crc kubenswrapper[4809]: E1205 11:09:59.872532 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:09:59 crc kubenswrapper[4809]: E1205 11:09:59.872942 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:09:59 crc kubenswrapper[4809]: E1205 11:09:59.873026 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.920196 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.920238 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.920251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.920271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.920286 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.931359 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.931388 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.931401 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.931416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.931425 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: E1205 11:09:59.946260 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.950903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.950942 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.950953 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.950973 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.950985 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: E1205 11:09:59.965370 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.971773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.971840 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.971853 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.971880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.971896 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:09:59 crc kubenswrapper[4809]: E1205 11:09:59.985739 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:09:59Z is after 2025-08-24T17:21:41Z" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.990806 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.990863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.990877 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.990929 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:09:59 crc kubenswrapper[4809]: I1205 11:09:59.990944 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:09:59Z","lastTransitionTime":"2025-12-05T11:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: E1205 11:10:00.003542 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:10:00Z is after 2025-08-24T17:21:41Z" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.008452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.008495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.008507 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.008525 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.008535 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: E1205 11:10:00.030919 4809 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T11:10:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2141eb26-0a7e-4c21-87d5-ee9ba8842777\\\",\\\"systemUUID\\\":\\\"67643894-7483-44e6-a6f9-98b40195f076\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T11:10:00Z is after 2025-08-24T17:21:41Z" Dec 05 11:10:00 crc kubenswrapper[4809]: E1205 11:10:00.031084 4809 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.033092 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.033125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.033135 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.033154 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.033165 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.135390 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.135436 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.135445 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.135459 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.135467 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.237343 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.237378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.237395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.237413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.237620 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.341402 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.341741 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.341796 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.341831 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.341854 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.444901 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.444965 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.444979 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.445000 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.445014 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.548280 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.548344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.548365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.548400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.548425 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.650720 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.650785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.650797 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.650814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.650826 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.753078 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.753122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.753134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.753152 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.753161 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.855482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.855520 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.855529 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.855541 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.855550 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.958096 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.958178 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.958190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.958207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:00 crc kubenswrapper[4809]: I1205 11:10:00.958220 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:00Z","lastTransitionTime":"2025-12-05T11:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.060371 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.060427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.060444 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.060468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.060488 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.163021 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.163060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.163068 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.163083 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.163092 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.266365 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.266400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.266408 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.266424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.266434 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.369596 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.369704 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.369728 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.369757 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.369780 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.473236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.473288 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.473303 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.473323 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.473337 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.575992 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.576039 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.576053 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.576073 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.576084 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.679137 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.679231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.679251 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.679286 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.679306 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.782729 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.782807 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.782823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.782848 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.782864 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.872148 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.872243 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:01 crc kubenswrapper[4809]: E1205 11:10:01.872336 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.872262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:01 crc kubenswrapper[4809]: E1205 11:10:01.872437 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:01 crc kubenswrapper[4809]: E1205 11:10:01.872566 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.872751 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:01 crc kubenswrapper[4809]: E1205 11:10:01.872834 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.885049 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.885089 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.885099 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.885117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.885128 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.987475 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.987535 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.987553 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.987574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:01 crc kubenswrapper[4809]: I1205 11:10:01.987586 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:01Z","lastTransitionTime":"2025-12-05T11:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.091122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.091174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.091190 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.091216 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.091233 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.194306 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.194361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.194372 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.194398 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.194410 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.297801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.297866 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.297878 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.297894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.297906 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.406292 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.406339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.406352 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.406374 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.406387 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.510744 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.510817 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.510841 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.510872 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.510896 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.614020 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.614087 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.614110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.614141 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.614163 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.716903 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.716974 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.717002 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.717030 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.717049 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.820086 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.820153 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.820177 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.820207 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.820229 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.923416 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.923486 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.923511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.923540 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:02 crc kubenswrapper[4809]: I1205 11:10:02.923560 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:02Z","lastTransitionTime":"2025-12-05T11:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.027198 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.027287 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.027386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.027413 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.027435 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.131271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.131361 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.131386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.131419 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.131440 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.234332 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.234382 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.234400 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.234427 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.234445 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.336785 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.336827 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.336839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.336859 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.336872 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.440124 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.440204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.440217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.440237 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.440251 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.543572 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.543609 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.543620 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.543670 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.543685 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.646168 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.646417 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.646472 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.646501 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.646515 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.749344 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.749384 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.749395 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.749418 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.749433 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.852354 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.852452 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.852473 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.852506 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.852531 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.871909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:03 crc kubenswrapper[4809]: E1205 11:10:03.872055 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.872158 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.872272 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.872178 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:03 crc kubenswrapper[4809]: E1205 11:10:03.872385 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:03 crc kubenswrapper[4809]: E1205 11:10:03.872508 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:03 crc kubenswrapper[4809]: E1205 11:10:03.872691 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.955570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.955627 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.955662 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.955681 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:03 crc kubenswrapper[4809]: I1205 11:10:03.955694 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:03Z","lastTransitionTime":"2025-12-05T11:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.059536 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.059653 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.059665 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.059689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.059707 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.162782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.162846 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.162863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.162887 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.162904 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.265483 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.265552 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.265570 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.265595 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.265617 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.368825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.368880 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.368893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.368919 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.368935 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.472117 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.472171 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.472182 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.472204 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.472219 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.575429 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.575482 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.575495 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.575518 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.575531 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.679172 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.679231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.679245 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.679265 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.679276 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.782393 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.782457 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.782468 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.782489 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.782502 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.871991 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:10:04 crc kubenswrapper[4809]: E1205 11:10:04.872122 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.884693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.884733 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.884742 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.884756 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.884767 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.892757 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ssnzv" podStartSLOduration=76.892712626 podStartE2EDuration="1m16.892712626s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:04.892047515 +0000 UTC m=+100.283024083" watchObservedRunningTime="2025-12-05 11:10:04.892712626 +0000 UTC m=+100.283689184" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.971786 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rjsd4" podStartSLOduration=76.971769041 podStartE2EDuration="1m16.971769041s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:04.970789479 +0000 UTC m=+100.361766057" watchObservedRunningTime="2025-12-05 11:10:04.971769041 +0000 UTC m=+100.362745599" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.987415 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.987453 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.987464 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.987480 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:04 crc kubenswrapper[4809]: I1205 11:10:04.987492 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:04Z","lastTransitionTime":"2025-12-05T11:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.018219 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-l5kgl" podStartSLOduration=77.018202508 podStartE2EDuration="1m17.018202508s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.003601893 +0000 UTC m=+100.394578471" watchObservedRunningTime="2025-12-05 11:10:05.018202508 +0000 UTC m=+100.409179056" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.018462 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=51.018456696 podStartE2EDuration="51.018456696s" podCreationTimestamp="2025-12-05 11:09:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.017942509 +0000 UTC m=+100.408919067" watchObservedRunningTime="2025-12-05 11:10:05.018456696 +0000 UTC m=+100.409433254" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.060443 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=83.060425471 podStartE2EDuration="1m23.060425471s" podCreationTimestamp="2025-12-05 11:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.050832775 +0000 UTC m=+100.441809383" watchObservedRunningTime="2025-12-05 11:10:05.060425471 +0000 UTC m=+100.451402039" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.076313 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zwhnj" podStartSLOduration=77.076294465 podStartE2EDuration="1m17.076294465s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.075292994 +0000 UTC m=+100.466269552" watchObservedRunningTime="2025-12-05 11:10:05.076294465 +0000 UTC m=+100.467271023" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.089657 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.089706 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.089716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.089734 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.089745 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.092104 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podStartSLOduration=77.092087928 podStartE2EDuration="1m17.092087928s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.092034976 +0000 UTC m=+100.483011544" watchObservedRunningTime="2025-12-05 11:10:05.092087928 +0000 UTC m=+100.483064486" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.105186 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-n79nb" podStartSLOduration=77.105172364 podStartE2EDuration="1m17.105172364s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.1050502 +0000 UTC m=+100.496026768" watchObservedRunningTime="2025-12-05 11:10:05.105172364 +0000 UTC m=+100.496148942" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.121498 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=83.121471582 podStartE2EDuration="1m23.121471582s" podCreationTimestamp="2025-12-05 11:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.120667287 +0000 UTC m=+100.511643865" watchObservedRunningTime="2025-12-05 11:10:05.121471582 +0000 UTC m=+100.512448140" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.158710 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=81.158690286 podStartE2EDuration="1m21.158690286s" podCreationTimestamp="2025-12-05 11:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.146287382 +0000 UTC m=+100.537263950" watchObservedRunningTime="2025-12-05 11:10:05.158690286 +0000 UTC m=+100.549666854" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.159289 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=29.159278335 podStartE2EDuration="29.159278335s" podCreationTimestamp="2025-12-05 11:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:05.158395267 +0000 UTC m=+100.549371825" watchObservedRunningTime="2025-12-05 11:10:05.159278335 +0000 UTC m=+100.550254913" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.191893 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.191939 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.191950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.191967 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.191978 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.295110 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.295193 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.295217 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.295247 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.295272 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.397719 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.397772 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.397790 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.397814 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.397830 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.500469 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.500511 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.500522 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.500537 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.500550 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.603910 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.603964 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.603983 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.604006 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.604023 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.707705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.707762 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.707779 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.707801 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.707818 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.809769 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.809810 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.809823 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.809842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.809854 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.871877 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.871993 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.872192 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:05 crc kubenswrapper[4809]: E1205 11:10:05.872157 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.872231 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:05 crc kubenswrapper[4809]: E1205 11:10:05.872340 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:05 crc kubenswrapper[4809]: E1205 11:10:05.872385 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:05 crc kubenswrapper[4809]: E1205 11:10:05.872448 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.911896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.911927 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.911936 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.911950 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:05 crc kubenswrapper[4809]: I1205 11:10:05.911960 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:05Z","lastTransitionTime":"2025-12-05T11:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.014308 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.014348 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.014358 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.014373 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.014382 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.117326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.117366 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.117378 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.117396 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.117409 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.219926 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.219968 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.219977 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.219991 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.219999 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.323326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.323389 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.323403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.323424 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.323441 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.426044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.426122 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.426142 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.426169 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.426185 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.528061 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.528097 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.528108 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.528123 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.528137 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.630285 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.630330 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.630339 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.630355 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.630368 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.732543 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.732594 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.732608 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.732650 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.732668 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.836225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.836842 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.836863 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.836884 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.836901 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.942824 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.943222 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.943246 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.943260 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:06 crc kubenswrapper[4809]: I1205 11:10:06.943269 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:06Z","lastTransitionTime":"2025-12-05T11:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.045996 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.046035 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.046046 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.046060 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.046072 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.148775 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.148818 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.148828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.148845 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.148856 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.250621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.250689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.250699 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.250716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.250726 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.353231 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.353273 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.353284 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.353300 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.353310 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.455592 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.455663 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.455673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.455691 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.455701 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.558183 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.558215 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.558225 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.558242 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.558253 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.661044 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.661088 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.661101 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.661116 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.661125 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.764070 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.764111 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.764120 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.764134 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.764144 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.866976 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.867326 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.867340 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.867357 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.867369 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.871283 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.871283 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.871417 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:07 crc kubenswrapper[4809]: E1205 11:10:07.871579 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.871596 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:07 crc kubenswrapper[4809]: E1205 11:10:07.871736 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:07 crc kubenswrapper[4809]: E1205 11:10:07.871818 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:07 crc kubenswrapper[4809]: E1205 11:10:07.871853 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.969275 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.969324 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.969334 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.969351 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.969362 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:07Z","lastTransitionTime":"2025-12-05T11:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:07 crc kubenswrapper[4809]: I1205 11:10:07.990950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:07 crc kubenswrapper[4809]: E1205 11:10:07.991105 4809 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:10:07 crc kubenswrapper[4809]: E1205 11:10:07.991173 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs podName:700689a7-7e7b-473c-8f2d-cfdd43d8e8f9 nodeName:}" failed. No retries permitted until 2025-12-05 11:11:11.991153232 +0000 UTC m=+167.382129790 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs") pod "network-metrics-daemon-xxkn4" (UID: "700689a7-7e7b-473c-8f2d-cfdd43d8e8f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.071497 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.071538 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.071547 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.071563 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.071573 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.174505 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.174559 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.174574 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.174597 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.174610 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.277828 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.277894 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.277911 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.277934 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.277953 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.380612 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.380722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.380754 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.380773 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.380785 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.483174 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.483236 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.483249 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.483266 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.483277 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.585782 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.585825 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.585839 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.585855 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.585866 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.688621 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.688725 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.688738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.688755 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.688768 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.791545 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.791678 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.791693 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.791716 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.791753 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.894544 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.894589 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.894599 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.894614 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.894625 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.997298 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.997335 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.997342 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.997356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:08 crc kubenswrapper[4809]: I1205 11:10:08.997364 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:08Z","lastTransitionTime":"2025-12-05T11:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.100312 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.100356 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.100386 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.100406 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.100425 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.202618 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.202710 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.202726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.202746 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.202762 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.304578 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.304652 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.304664 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.304682 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.304697 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.407329 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.407376 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.407387 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.407403 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.407415 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.510651 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.511470 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.511493 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.511513 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.511526 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.613896 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.613935 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.613946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.613959 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.613968 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.722437 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.722689 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.722705 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.722722 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.722734 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.825891 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.825935 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.825946 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.825961 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.825975 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.871276 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:09 crc kubenswrapper[4809]: E1205 11:10:09.871417 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.871610 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:09 crc kubenswrapper[4809]: E1205 11:10:09.871700 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.871830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.871864 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:09 crc kubenswrapper[4809]: E1205 11:10:09.871899 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:09 crc kubenswrapper[4809]: E1205 11:10:09.872015 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.928565 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.928656 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.928668 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.928688 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:09 crc kubenswrapper[4809]: I1205 11:10:09.928701 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:09Z","lastTransitionTime":"2025-12-05T11:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.031065 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.031113 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.031125 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.031144 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.031157 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:10Z","lastTransitionTime":"2025-12-05T11:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.133383 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.133410 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.133421 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.133435 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.133444 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:10Z","lastTransitionTime":"2025-12-05T11:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.236230 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.236271 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.236281 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.236299 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.236312 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:10Z","lastTransitionTime":"2025-12-05T11:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.340554 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.340603 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.340616 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.340647 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.340799 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:10Z","lastTransitionTime":"2025-12-05T11:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.357673 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.357717 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.357726 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.357738 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.357746 4809 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T11:10:10Z","lastTransitionTime":"2025-12-05T11:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.400848 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j"] Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.401271 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.403718 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.404305 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.404793 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.412798 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.515418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c52f678f-e02a-4439-b640-231395bb24bd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.515475 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c52f678f-e02a-4439-b640-231395bb24bd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.515502 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52f678f-e02a-4439-b640-231395bb24bd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.515522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c52f678f-e02a-4439-b640-231395bb24bd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.515619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c52f678f-e02a-4439-b640-231395bb24bd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.616948 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c52f678f-e02a-4439-b640-231395bb24bd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.616994 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c52f678f-e02a-4439-b640-231395bb24bd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.617016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c52f678f-e02a-4439-b640-231395bb24bd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.617038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52f678f-e02a-4439-b640-231395bb24bd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.617059 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c52f678f-e02a-4439-b640-231395bb24bd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.617109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c52f678f-e02a-4439-b640-231395bb24bd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.617190 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c52f678f-e02a-4439-b640-231395bb24bd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.618218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c52f678f-e02a-4439-b640-231395bb24bd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.625771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c52f678f-e02a-4439-b640-231395bb24bd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.636979 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c52f678f-e02a-4439-b640-231395bb24bd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4z94j\" (UID: \"c52f678f-e02a-4439-b640-231395bb24bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: I1205 11:10:10.717906 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" Dec 05 11:10:10 crc kubenswrapper[4809]: W1205 11:10:10.733872 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc52f678f_e02a_4439_b640_231395bb24bd.slice/crio-4c4308e6ca214cdcd5949f8c2c80ae2c8e6c94764a14708954895e749f7e93e4 WatchSource:0}: Error finding container 4c4308e6ca214cdcd5949f8c2c80ae2c8e6c94764a14708954895e749f7e93e4: Status 404 returned error can't find the container with id 4c4308e6ca214cdcd5949f8c2c80ae2c8e6c94764a14708954895e749f7e93e4 Dec 05 11:10:11 crc kubenswrapper[4809]: I1205 11:10:11.492165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" event={"ID":"c52f678f-e02a-4439-b640-231395bb24bd","Type":"ContainerStarted","Data":"6a76fe7649ab4fbbda697f07033cc79483c61eb1a7c7af99a44f0c1d6e3bd52e"} Dec 05 11:10:11 crc kubenswrapper[4809]: I1205 11:10:11.492212 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" event={"ID":"c52f678f-e02a-4439-b640-231395bb24bd","Type":"ContainerStarted","Data":"4c4308e6ca214cdcd5949f8c2c80ae2c8e6c94764a14708954895e749f7e93e4"} Dec 05 11:10:11 crc kubenswrapper[4809]: I1205 11:10:11.872186 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:11 crc kubenswrapper[4809]: I1205 11:10:11.872278 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:11 crc kubenswrapper[4809]: I1205 11:10:11.872276 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:11 crc kubenswrapper[4809]: E1205 11:10:11.872738 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:11 crc kubenswrapper[4809]: E1205 11:10:11.872788 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:11 crc kubenswrapper[4809]: I1205 11:10:11.872318 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:11 crc kubenswrapper[4809]: E1205 11:10:11.872855 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:11 crc kubenswrapper[4809]: E1205 11:10:11.873006 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:13 crc kubenswrapper[4809]: I1205 11:10:13.872074 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:13 crc kubenswrapper[4809]: I1205 11:10:13.872120 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:13 crc kubenswrapper[4809]: I1205 11:10:13.872183 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:13 crc kubenswrapper[4809]: E1205 11:10:13.872300 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:13 crc kubenswrapper[4809]: I1205 11:10:13.872317 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:13 crc kubenswrapper[4809]: E1205 11:10:13.872428 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:13 crc kubenswrapper[4809]: E1205 11:10:13.872577 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:13 crc kubenswrapper[4809]: E1205 11:10:13.872719 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:15 crc kubenswrapper[4809]: I1205 11:10:15.871341 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:15 crc kubenswrapper[4809]: I1205 11:10:15.871341 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:15 crc kubenswrapper[4809]: I1205 11:10:15.871459 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:15 crc kubenswrapper[4809]: I1205 11:10:15.871700 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:15 crc kubenswrapper[4809]: E1205 11:10:15.871693 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:15 crc kubenswrapper[4809]: E1205 11:10:15.871808 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:15 crc kubenswrapper[4809]: E1205 11:10:15.871843 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:15 crc kubenswrapper[4809]: E1205 11:10:15.871921 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:17 crc kubenswrapper[4809]: I1205 11:10:17.871929 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:17 crc kubenswrapper[4809]: I1205 11:10:17.872014 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:17 crc kubenswrapper[4809]: E1205 11:10:17.872139 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:17 crc kubenswrapper[4809]: I1205 11:10:17.872185 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:17 crc kubenswrapper[4809]: I1205 11:10:17.872192 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:17 crc kubenswrapper[4809]: E1205 11:10:17.872290 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:17 crc kubenswrapper[4809]: E1205 11:10:17.872462 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:17 crc kubenswrapper[4809]: E1205 11:10:17.872585 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:17 crc kubenswrapper[4809]: I1205 11:10:17.873781 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:10:17 crc kubenswrapper[4809]: E1205 11:10:17.874022 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:10:19 crc kubenswrapper[4809]: I1205 11:10:19.871936 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:19 crc kubenswrapper[4809]: I1205 11:10:19.872001 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:19 crc kubenswrapper[4809]: I1205 11:10:19.871962 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:19 crc kubenswrapper[4809]: I1205 11:10:19.872096 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:19 crc kubenswrapper[4809]: E1205 11:10:19.872261 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:19 crc kubenswrapper[4809]: E1205 11:10:19.872533 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:19 crc kubenswrapper[4809]: E1205 11:10:19.872738 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:19 crc kubenswrapper[4809]: E1205 11:10:19.872897 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:21 crc kubenswrapper[4809]: I1205 11:10:21.871367 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:21 crc kubenswrapper[4809]: I1205 11:10:21.871469 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:21 crc kubenswrapper[4809]: I1205 11:10:21.871558 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:21 crc kubenswrapper[4809]: I1205 11:10:21.871589 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:21 crc kubenswrapper[4809]: E1205 11:10:21.871612 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:21 crc kubenswrapper[4809]: E1205 11:10:21.871756 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:21 crc kubenswrapper[4809]: E1205 11:10:21.871863 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:21 crc kubenswrapper[4809]: E1205 11:10:21.871979 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:23 crc kubenswrapper[4809]: I1205 11:10:23.872265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:23 crc kubenswrapper[4809]: I1205 11:10:23.872318 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:23 crc kubenswrapper[4809]: I1205 11:10:23.872265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:23 crc kubenswrapper[4809]: I1205 11:10:23.872475 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:23 crc kubenswrapper[4809]: E1205 11:10:23.872850 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:23 crc kubenswrapper[4809]: E1205 11:10:23.872929 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:23 crc kubenswrapper[4809]: E1205 11:10:23.872976 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:23 crc kubenswrapper[4809]: E1205 11:10:23.872767 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:24 crc kubenswrapper[4809]: E1205 11:10:24.843120 4809 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 05 11:10:25 crc kubenswrapper[4809]: E1205 11:10:25.073586 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.539214 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/1.log" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.539692 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/0.log" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.539743 4809 generic.go:334] "Generic (PLEG): container finished" podID="238d6f1b-f3b4-4b33-a33c-10a0c99e8178" containerID="65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf" exitCode=1 Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.539777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerDied","Data":"65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf"} Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.539818 4809 scope.go:117] "RemoveContainer" containerID="ff9c84881dbe2e7f2db7e169e6054354c2e5bc913df98fbd8333385bcbc0a4da" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.540425 4809 scope.go:117] "RemoveContainer" containerID="65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf" Dec 05 11:10:25 crc kubenswrapper[4809]: E1205 11:10:25.540751 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-n79nb_openshift-multus(238d6f1b-f3b4-4b33-a33c-10a0c99e8178)\"" pod="openshift-multus/multus-n79nb" podUID="238d6f1b-f3b4-4b33-a33c-10a0c99e8178" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.561866 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4z94j" podStartSLOduration=97.561838333 podStartE2EDuration="1m37.561838333s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:11.50487115 +0000 UTC m=+106.895847708" watchObservedRunningTime="2025-12-05 11:10:25.561838333 +0000 UTC m=+120.952814891" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.871402 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.871422 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.871463 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:25 crc kubenswrapper[4809]: E1205 11:10:25.872349 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:25 crc kubenswrapper[4809]: E1205 11:10:25.872145 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:25 crc kubenswrapper[4809]: E1205 11:10:25.872464 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:25 crc kubenswrapper[4809]: I1205 11:10:25.871525 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:25 crc kubenswrapper[4809]: E1205 11:10:25.872624 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:26 crc kubenswrapper[4809]: I1205 11:10:26.543840 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/1.log" Dec 05 11:10:27 crc kubenswrapper[4809]: I1205 11:10:27.871952 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:27 crc kubenswrapper[4809]: I1205 11:10:27.872028 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:27 crc kubenswrapper[4809]: I1205 11:10:27.871946 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:27 crc kubenswrapper[4809]: I1205 11:10:27.872100 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:27 crc kubenswrapper[4809]: E1205 11:10:27.872196 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:27 crc kubenswrapper[4809]: E1205 11:10:27.872307 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:27 crc kubenswrapper[4809]: E1205 11:10:27.872405 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:27 crc kubenswrapper[4809]: E1205 11:10:27.872567 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:29 crc kubenswrapper[4809]: I1205 11:10:29.871730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:29 crc kubenswrapper[4809]: E1205 11:10:29.872774 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:29 crc kubenswrapper[4809]: I1205 11:10:29.871786 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:29 crc kubenswrapper[4809]: E1205 11:10:29.873000 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:29 crc kubenswrapper[4809]: I1205 11:10:29.871736 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:29 crc kubenswrapper[4809]: E1205 11:10:29.873182 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:29 crc kubenswrapper[4809]: I1205 11:10:29.871807 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:29 crc kubenswrapper[4809]: E1205 11:10:29.873363 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:30 crc kubenswrapper[4809]: E1205 11:10:30.075656 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 11:10:30 crc kubenswrapper[4809]: I1205 11:10:30.873309 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:10:30 crc kubenswrapper[4809]: E1205 11:10:30.873709 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cdvk8_openshift-ovn-kubernetes(df22f1c5-aea5-4717-b2b0-3d6a577ea53f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" Dec 05 11:10:31 crc kubenswrapper[4809]: I1205 11:10:31.871811 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:31 crc kubenswrapper[4809]: I1205 11:10:31.871877 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:31 crc kubenswrapper[4809]: I1205 11:10:31.871982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:31 crc kubenswrapper[4809]: I1205 11:10:31.871865 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:31 crc kubenswrapper[4809]: E1205 11:10:31.872890 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:31 crc kubenswrapper[4809]: E1205 11:10:31.873485 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:31 crc kubenswrapper[4809]: E1205 11:10:31.873587 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:31 crc kubenswrapper[4809]: E1205 11:10:31.873367 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:33 crc kubenswrapper[4809]: I1205 11:10:33.871901 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:33 crc kubenswrapper[4809]: I1205 11:10:33.871941 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:33 crc kubenswrapper[4809]: E1205 11:10:33.872165 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:33 crc kubenswrapper[4809]: I1205 11:10:33.872196 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:33 crc kubenswrapper[4809]: I1205 11:10:33.872253 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:33 crc kubenswrapper[4809]: E1205 11:10:33.872381 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:33 crc kubenswrapper[4809]: E1205 11:10:33.872480 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:33 crc kubenswrapper[4809]: E1205 11:10:33.872567 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:35 crc kubenswrapper[4809]: E1205 11:10:35.077422 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 11:10:35 crc kubenswrapper[4809]: I1205 11:10:35.872138 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:35 crc kubenswrapper[4809]: I1205 11:10:35.872210 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:35 crc kubenswrapper[4809]: E1205 11:10:35.872275 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:35 crc kubenswrapper[4809]: I1205 11:10:35.872354 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:35 crc kubenswrapper[4809]: I1205 11:10:35.872559 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:35 crc kubenswrapper[4809]: E1205 11:10:35.872555 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:35 crc kubenswrapper[4809]: E1205 11:10:35.872612 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:35 crc kubenswrapper[4809]: E1205 11:10:35.872716 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:37 crc kubenswrapper[4809]: I1205 11:10:37.871419 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:37 crc kubenswrapper[4809]: I1205 11:10:37.871540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:37 crc kubenswrapper[4809]: I1205 11:10:37.871599 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:37 crc kubenswrapper[4809]: I1205 11:10:37.871618 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:37 crc kubenswrapper[4809]: E1205 11:10:37.871559 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:37 crc kubenswrapper[4809]: E1205 11:10:37.871777 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:37 crc kubenswrapper[4809]: E1205 11:10:37.871846 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:37 crc kubenswrapper[4809]: E1205 11:10:37.871919 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:39 crc kubenswrapper[4809]: I1205 11:10:39.871717 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:39 crc kubenswrapper[4809]: E1205 11:10:39.871846 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:39 crc kubenswrapper[4809]: I1205 11:10:39.871959 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:39 crc kubenswrapper[4809]: I1205 11:10:39.871991 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:39 crc kubenswrapper[4809]: E1205 11:10:39.872416 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:39 crc kubenswrapper[4809]: I1205 11:10:39.872475 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:39 crc kubenswrapper[4809]: E1205 11:10:39.872486 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:39 crc kubenswrapper[4809]: E1205 11:10:39.872628 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:39 crc kubenswrapper[4809]: I1205 11:10:39.872995 4809 scope.go:117] "RemoveContainer" containerID="65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf" Dec 05 11:10:40 crc kubenswrapper[4809]: E1205 11:10:40.079154 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 11:10:40 crc kubenswrapper[4809]: I1205 11:10:40.596383 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/1.log" Dec 05 11:10:40 crc kubenswrapper[4809]: I1205 11:10:40.596841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerStarted","Data":"5861e92534d270984c5ea325ee01014773b774019a667ed88e3490f52b54c167"} Dec 05 11:10:41 crc kubenswrapper[4809]: I1205 11:10:41.871548 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:41 crc kubenswrapper[4809]: I1205 11:10:41.871623 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:41 crc kubenswrapper[4809]: I1205 11:10:41.871665 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:41 crc kubenswrapper[4809]: I1205 11:10:41.871574 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:41 crc kubenswrapper[4809]: E1205 11:10:41.871846 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:41 crc kubenswrapper[4809]: E1205 11:10:41.871967 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:41 crc kubenswrapper[4809]: E1205 11:10:41.872784 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:41 crc kubenswrapper[4809]: E1205 11:10:41.872957 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:41 crc kubenswrapper[4809]: I1205 11:10:41.873788 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.613555 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/3.log" Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.616884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerStarted","Data":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.617715 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.650543 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podStartSLOduration=115.650523787 podStartE2EDuration="1m55.650523787s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:43.649659993 +0000 UTC m=+139.040636571" watchObservedRunningTime="2025-12-05 11:10:43.650523787 +0000 UTC m=+139.041500355" Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.871491 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.871587 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:43 crc kubenswrapper[4809]: E1205 11:10:43.871648 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.871659 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:43 crc kubenswrapper[4809]: E1205 11:10:43.871716 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:43 crc kubenswrapper[4809]: E1205 11:10:43.871810 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:43 crc kubenswrapper[4809]: I1205 11:10:43.871913 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:43 crc kubenswrapper[4809]: E1205 11:10:43.872030 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:44 crc kubenswrapper[4809]: I1205 11:10:44.073215 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xxkn4"] Dec 05 11:10:44 crc kubenswrapper[4809]: I1205 11:10:44.619838 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:44 crc kubenswrapper[4809]: E1205 11:10:44.620287 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:45 crc kubenswrapper[4809]: E1205 11:10:45.080219 4809 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 11:10:45 crc kubenswrapper[4809]: I1205 11:10:45.871998 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:45 crc kubenswrapper[4809]: I1205 11:10:45.872090 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:45 crc kubenswrapper[4809]: E1205 11:10:45.873446 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:45 crc kubenswrapper[4809]: I1205 11:10:45.872173 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:45 crc kubenswrapper[4809]: E1205 11:10:45.873710 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:45 crc kubenswrapper[4809]: E1205 11:10:45.873780 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:46 crc kubenswrapper[4809]: I1205 11:10:46.871887 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:46 crc kubenswrapper[4809]: E1205 11:10:46.872023 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:47 crc kubenswrapper[4809]: I1205 11:10:47.871283 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:47 crc kubenswrapper[4809]: E1205 11:10:47.871398 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:47 crc kubenswrapper[4809]: I1205 11:10:47.871404 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:47 crc kubenswrapper[4809]: E1205 11:10:47.871597 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:47 crc kubenswrapper[4809]: I1205 11:10:47.871619 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:47 crc kubenswrapper[4809]: E1205 11:10:47.871706 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:48 crc kubenswrapper[4809]: I1205 11:10:48.872203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:48 crc kubenswrapper[4809]: E1205 11:10:48.872341 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxkn4" podUID="700689a7-7e7b-473c-8f2d-cfdd43d8e8f9" Dec 05 11:10:49 crc kubenswrapper[4809]: I1205 11:10:49.871662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:49 crc kubenswrapper[4809]: E1205 11:10:49.871795 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 11:10:49 crc kubenswrapper[4809]: I1205 11:10:49.871993 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:49 crc kubenswrapper[4809]: E1205 11:10:49.872079 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 11:10:49 crc kubenswrapper[4809]: I1205 11:10:49.872361 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:49 crc kubenswrapper[4809]: E1205 11:10:49.872704 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 11:10:50 crc kubenswrapper[4809]: I1205 11:10:50.872082 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:10:50 crc kubenswrapper[4809]: I1205 11:10:50.874965 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 11:10:50 crc kubenswrapper[4809]: I1205 11:10:50.875571 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.093498 4809 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.124489 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kpt8q"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.125086 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.129261 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.130057 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.130619 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.130737 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gvwf"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.131178 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.131320 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.134518 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f8b2x"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.135170 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.142413 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.142930 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.149417 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.149648 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.150140 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.151036 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.151271 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.154288 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.154620 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.157900 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.159770 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.171430 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.172031 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.172085 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.172720 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.173379 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.173510 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.173680 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.173798 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.173862 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hhcsc"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.173897 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.173916 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174009 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174094 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174115 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174145 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174683 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174693 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174839 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174903 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.174934 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175004 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175144 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175169 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175307 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175382 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175386 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175443 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175502 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175520 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175604 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175687 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175745 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175789 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175848 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.175995 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.176089 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.176190 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.176288 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.176420 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.180805 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.181224 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.181459 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.181470 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.181514 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.181741 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.181956 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.182389 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.182410 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.182699 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.182855 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-shlh8"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.183179 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.183222 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.183467 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.183536 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.183764 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.183984 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.184110 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.184132 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.184565 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.184602 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.184896 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6vmbb"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.185216 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.186299 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.186934 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.188978 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6fvxb"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.189239 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.190025 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zhtzr"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.190134 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.190289 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.190499 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.190579 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4hqr"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.190918 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.191677 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.191736 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.191691 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.192406 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.192409 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.192683 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.193380 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f8b2x"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.193931 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.194058 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.194098 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.194165 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.194317 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.194320 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.194928 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.195101 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.195168 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.195201 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.195279 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.195510 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.202049 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.202563 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.203108 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.203689 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.203757 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76"] Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.204169 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.204169 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.204571 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 11:10:51 crc kubenswrapper[4809]: I1205 11:10:51.205475 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.400608 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.401257 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.401362 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.401397 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.401663 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.401768 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.401915 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.401959 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.402010 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.402271 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.402483 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.402540 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.402558 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.402607 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.403793 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zhtzr"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.403940 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.403966 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.404771 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-client-ca\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.404838 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.404905 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-config\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.404939 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4j85\" (UniqueName: \"kubernetes.io/projected/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-kube-api-access-f4j85\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405003 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405049 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s45t\" (UniqueName: \"kubernetes.io/projected/2ddee1c1-dd55-429b-b277-214aa6556913-kube-api-access-5s45t\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405103 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-config\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405128 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6rxg\" (UniqueName: \"kubernetes.io/projected/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-kube-api-access-b6rxg\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405154 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b52152a3-7674-489f-b67b-e96346e88228-serving-cert\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: E1205 11:10:52.405288 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:12:54.405268768 +0000 UTC m=+269.796245406 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405350 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-service-ca-bundle\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405377 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz7v5\" (UniqueName: \"kubernetes.io/projected/3c896868-d020-4367-aad2-9497223c6026-kube-api-access-qz7v5\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.405519 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-encryption-config\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406526 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hhcsc"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406649 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-etcd-client\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406682 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/73360c2b-d090-42c2-8800-94b0700ae627-audit-dir\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406728 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-etcd-serving-ca\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-image-import-ca\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406820 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406853 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406877 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-etcd-client\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsmlv\" (UniqueName: \"kubernetes.io/projected/27d1916f-dfbe-4e36-963c-958803f4eb66-kube-api-access-dsmlv\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc69z\" (UniqueName: \"kubernetes.io/projected/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-kube-api-access-bc69z\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-images\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406967 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ddee1c1-dd55-429b-b277-214aa6556913-auth-proxy-config\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.406986 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-config\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407119 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddee1c1-dd55-429b-b277-214aa6556913-config\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407216 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-audit-policies\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqqg9\" (UniqueName: \"kubernetes.io/projected/73360c2b-d090-42c2-8800-94b0700ae627-kube-api-access-jqqg9\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407353 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407386 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmr6j\" (UniqueName: \"kubernetes.io/projected/0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978-kube-api-access-qmr6j\") pod \"cluster-samples-operator-665b6dd947-df5wd\" (UID: \"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-config\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-audit\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-node-pullsecrets\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-encryption-config\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c896868-d020-4367-aad2-9497223c6026-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-serving-cert\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407590 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-audit-dir\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27d1916f-dfbe-4e36-963c-958803f4eb66-serving-cert\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8st7p\" (UniqueName: \"kubernetes.io/projected/b52152a3-7674-489f-b67b-e96346e88228-kube-api-access-8st7p\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-serving-cert\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407835 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c896868-d020-4367-aad2-9497223c6026-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-df5wd\" (UID: \"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407929 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ddee1c1-dd55-429b-b277-214aa6556913-machine-approver-tls\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.407973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.411373 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.412575 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.416058 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6vmbb"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.426928 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.427605 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.430841 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kpt8q"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.430543 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.430604 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.430686 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.430721 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.430755 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.430837 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.433359 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.431005 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.431057 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.431138 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.435325 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.438395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.441083 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.443499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.451393 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.455063 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.457313 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-shlh8"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.457371 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.460354 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.460392 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.460749 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.460833 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.462683 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.463261 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.463956 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qm6jb"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.464290 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-wml6l"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.474485 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.475013 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.475109 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.475458 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.475616 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.476295 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.476361 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.476993 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.478193 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.478419 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.479420 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.479975 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx7gs"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.481918 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.482003 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.482026 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.482725 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.483321 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.484969 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.487129 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.487289 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.498450 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.499740 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.499960 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.501207 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.501474 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.502266 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.502716 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-xgftq"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.503112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.508670 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.508907 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509039 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509186 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509270 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509042 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509490 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509800 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509845 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.509932 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.510044 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.511751 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512095 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-serving-cert\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512165 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ddee1c1-dd55-429b-b277-214aa6556913-machine-approver-tls\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512218 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c896868-d020-4367-aad2-9497223c6026-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-df5wd\" (UID: \"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512267 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-trusted-ca-bundle\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512285 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a80826b-26aa-45d4-991f-d1b585ab1a49-config\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-config\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl4z9\" (UniqueName: \"kubernetes.io/projected/677864f6-90af-4ffa-9b36-729194cedcd5-kube-api-access-jl4z9\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512368 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4j85\" (UniqueName: \"kubernetes.io/projected/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-kube-api-access-f4j85\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512432 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-client-ca\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512471 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-config\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512495 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512527 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-policies\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512551 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ntw8\" (UniqueName: \"kubernetes.io/projected/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-kube-api-access-2ntw8\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-config\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512608 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8f413f0a-b891-46aa-975f-388300b1931d-metrics-tls\") pod \"dns-operator-744455d44c-shlh8\" (UID: \"8f413f0a-b891-46aa-975f-388300b1931d\") " pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512669 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s45t\" (UniqueName: \"kubernetes.io/projected/2ddee1c1-dd55-429b-b277-214aa6556913-kube-api-access-5s45t\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512690 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6rxg\" (UniqueName: \"kubernetes.io/projected/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-kube-api-access-b6rxg\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512731 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-serving-cert\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512749 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9371528e-0a6d-4946-bb4f-e4781e098460-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512768 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-trusted-ca\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512789 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b52152a3-7674-489f-b67b-e96346e88228-serving-cert\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512808 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-service-ca-bundle\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-encryption-config\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512843 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-client-ca\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512864 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqdx2\" (UniqueName: \"kubernetes.io/projected/8f413f0a-b891-46aa-975f-388300b1931d-kube-api-access-dqdx2\") pod \"dns-operator-744455d44c-shlh8\" (UID: \"8f413f0a-b891-46aa-975f-388300b1931d\") " pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz7v5\" (UniqueName: \"kubernetes.io/projected/3c896868-d020-4367-aad2-9497223c6026-kube-api-access-qz7v5\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512909 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9371528e-0a6d-4946-bb4f-e4781e098460-config\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512928 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-etcd-client\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512949 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/73360c2b-d090-42c2-8800-94b0700ae627-audit-dir\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512968 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.512991 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513009 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513068 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-metrics-tls\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513100 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-etcd-serving-ca\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513119 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-service-ca\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513123 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c896868-d020-4367-aad2-9497223c6026-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513140 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b7406036-f795-40ba-b663-6b17adeabc8a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a80826b-26aa-45d4-991f-d1b585ab1a49-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513170 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-image-import-ca\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513240 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-etcd-client\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsmlv\" (UniqueName: \"kubernetes.io/projected/27d1916f-dfbe-4e36-963c-958803f4eb66-kube-api-access-dsmlv\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc69z\" (UniqueName: \"kubernetes.io/projected/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-kube-api-access-bc69z\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513438 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-images\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513459 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513478 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmhtr\" (UniqueName: \"kubernetes.io/projected/c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4-kube-api-access-rmhtr\") pod \"downloads-7954f5f757-6fvxb\" (UID: \"c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4\") " pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513498 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513518 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/86de2a01-5e19-41e5-ba51-6acd28109d9c-proxy-tls\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513540 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ddee1c1-dd55-429b-b277-214aa6556913-auth-proxy-config\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513560 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-config\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-serving-cert\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513646 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddee1c1-dd55-429b-b277-214aa6556913-config\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-serving-cert\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513686 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a80826b-26aa-45d4-991f-d1b585ab1a49-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-dir\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513727 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/86de2a01-5e19-41e5-ba51-6acd28109d9c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.513750 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-audit-policies\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.514020 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.516593 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-image-import-ca\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.516929 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518313 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518420 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-config\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518481 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqqg9\" (UniqueName: \"kubernetes.io/projected/73360c2b-d090-42c2-8800-94b0700ae627-kube-api-access-jqqg9\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmr6j\" (UniqueName: \"kubernetes.io/projected/0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978-kube-api-access-qmr6j\") pod \"cluster-samples-operator-665b6dd947-df5wd\" (UID: \"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518536 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7406036-f795-40ba-b663-6b17adeabc8a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-config-volume\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518643 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv8gf\" (UniqueName: \"kubernetes.io/projected/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-kube-api-access-fv8gf\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518683 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-config\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/86de2a01-5e19-41e5-ba51-6acd28109d9c-images\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518735 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqvgv\" (UniqueName: \"kubernetes.io/projected/86de2a01-5e19-41e5-ba51-6acd28109d9c-kube-api-access-zqvgv\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518767 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8pmc\" (UniqueName: \"kubernetes.io/projected/b7406036-f795-40ba-b663-6b17adeabc8a-kube-api-access-b8pmc\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-oauth-config\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518850 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7406036-f795-40ba-b663-6b17adeabc8a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-audit\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9371528e-0a6d-4946-bb4f-e4781e098460-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518931 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-node-pullsecrets\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-encryption-config\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.518977 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-oauth-serving-cert\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519002 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6p4v\" (UniqueName: \"kubernetes.io/projected/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-kube-api-access-z6p4v\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-config\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9fzm\" (UniqueName: \"kubernetes.io/projected/41d11685-f01b-4304-a08e-3c43f21e1411-kube-api-access-d9fzm\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519089 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519118 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-serving-cert\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519164 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kb6s\" (UniqueName: \"kubernetes.io/projected/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-kube-api-access-7kb6s\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c896868-d020-4367-aad2-9497223c6026-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519335 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519369 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-audit-dir\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-audit-dir\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519458 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27d1916f-dfbe-4e36-963c-958803f4eb66-serving-cert\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519505 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-console-config\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8st7p\" (UniqueName: \"kubernetes.io/projected/b52152a3-7674-489f-b67b-e96346e88228-kube-api-access-8st7p\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519569 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677864f6-90af-4ffa-9b36-729194cedcd5-serving-cert\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.519986 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-etcd-serving-ca\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.520522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.521182 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-config\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.521230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddee1c1-dd55-429b-b277-214aa6556913-config\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.521297 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/73360c2b-d090-42c2-8800-94b0700ae627-audit-dir\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.521755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.521978 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-service-ca-bundle\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.522255 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-images\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.522762 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-serving-cert\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.522996 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27d1916f-dfbe-4e36-963c-958803f4eb66-serving-cert\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.523161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/27d1916f-dfbe-4e36-963c-958803f4eb66-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.523247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-audit\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.523353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-client-ca\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.523547 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-node-pullsecrets\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.524142 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.524359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.524779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b52152a3-7674-489f-b67b-e96346e88228-serving-cert\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.524020 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ddee1c1-dd55-429b-b277-214aa6556913-auth-proxy-config\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.526012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/73360c2b-d090-42c2-8800-94b0700ae627-audit-policies\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.514528 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.527820 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.528497 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.529006 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.530569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-df5wd\" (UID: \"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.531709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.532493 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.533359 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.534180 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-config\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.534901 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-8d57d"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.535131 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-config\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.535582 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536115 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536215 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-serving-cert\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536320 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536491 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536524 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536622 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-etcd-client\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536586 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.536649 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.538876 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.539664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c896868-d020-4367-aad2-9497223c6026-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.539792 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqqg9\" (UniqueName: \"kubernetes.io/projected/73360c2b-d090-42c2-8800-94b0700ae627-kube-api-access-jqqg9\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.540074 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.540252 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.541050 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.541483 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2rpwq"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.541552 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/73360c2b-d090-42c2-8800-94b0700ae627-encryption-config\") pod \"apiserver-7bbb656c7d-w7ffm\" (UID: \"73360c2b-d090-42c2-8800-94b0700ae627\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.541807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsmlv\" (UniqueName: \"kubernetes.io/projected/27d1916f-dfbe-4e36-963c-958803f4eb66-kube-api-access-dsmlv\") pod \"authentication-operator-69f744f599-f8b2x\" (UID: \"27d1916f-dfbe-4e36-963c-958803f4eb66\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.542256 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-encryption-config\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.542323 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pbsgg"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.542525 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.542755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8st7p\" (UniqueName: \"kubernetes.io/projected/b52152a3-7674-489f-b67b-e96346e88228-kube-api-access-8st7p\") pod \"controller-manager-879f6c89f-7gvwf\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.542799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6rxg\" (UniqueName: \"kubernetes.io/projected/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-kube-api-access-b6rxg\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.542998 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-etcd-client\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.543182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.544082 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.544261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmr6j\" (UniqueName: \"kubernetes.io/projected/0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978-kube-api-access-qmr6j\") pod \"cluster-samples-operator-665b6dd947-df5wd\" (UID: \"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.544366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2ddee1c1-dd55-429b-b277-214aa6556913-machine-approver-tls\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.546085 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fceeabf7-b542-4506-a6c2-dae29b3ad4a3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7l888\" (UID: \"fceeabf7-b542-4506-a6c2-dae29b3ad4a3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.546152 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.547375 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wfbxx"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.547781 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s45t\" (UniqueName: \"kubernetes.io/projected/2ddee1c1-dd55-429b-b277-214aa6556913-kube-api-access-5s45t\") pod \"machine-approver-56656f9798-gcwbn\" (UID: \"2ddee1c1-dd55-429b-b277-214aa6556913\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.548013 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.548555 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.551267 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4hqr"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.552086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc69z\" (UniqueName: \"kubernetes.io/projected/9cf0c441-26db-4ac8-a1ed-4c2c90bc214c-kube-api-access-bc69z\") pod \"machine-api-operator-5694c8668f-6vmbb\" (UID: \"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.552747 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gvwf"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.554312 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9b4bq"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.555273 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.559030 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.559613 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.560121 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-cv89t"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.560561 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.560730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.561019 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.563435 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.563978 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.566723 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.567378 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.568119 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.569025 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.569113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.569288 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.573721 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6fvxb"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.574201 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.575258 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.576360 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qm6jb"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.582672 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx7gs"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.582720 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.582731 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.586782 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.586844 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9b4bq"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.588427 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.590565 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2rpwq"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.592695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.593685 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wml6l"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.594026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4j85\" (UniqueName: \"kubernetes.io/projected/d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2-kube-api-access-f4j85\") pod \"apiserver-76f77b778f-kpt8q\" (UID: \"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2\") " pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.598699 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.598748 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.600325 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.602697 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pbsgg"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.604459 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.605640 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.607262 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.607577 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.608425 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz7v5\" (UniqueName: \"kubernetes.io/projected/3c896868-d020-4367-aad2-9497223c6026-kube-api-access-qz7v5\") pod \"openshift-controller-manager-operator-756b6f6bc6-8fcv4\" (UID: \"3c896868-d020-4367-aad2-9497223c6026\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.608657 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.610183 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.611581 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wfbxx"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.613194 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.614403 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8d57d"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.615649 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.620796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.620846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-tls\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.620876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-trusted-ca-bundle\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.620898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-config\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.620918 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-bound-sa-token\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.620947 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.620974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-default-certificate\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621045 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9371528e-0a6d-4946-bb4f-e4781e098460-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621087 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-config\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-serving-cert\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621236 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-trusted-ca\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621262 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-ca\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621294 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9371528e-0a6d-4946-bb4f-e4781e098460-config\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621320 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-metrics-tls\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-service-ca\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-serving-cert\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621418 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b7406036-f795-40ba-b663-6b17adeabc8a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-trusted-ca\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1de639b4-f614-41b9-812b-a7bfea3ccc90-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-service-ca\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621536 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/86de2a01-5e19-41e5-ba51-6acd28109d9c-proxy-tls\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621562 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-serving-cert\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a80826b-26aa-45d4-991f-d1b585ab1a49-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621606 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-stats-auth\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-dir\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621679 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqvgv\" (UniqueName: \"kubernetes.io/projected/86de2a01-5e19-41e5-ba51-6acd28109d9c-kube-api-access-zqvgv\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621698 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv8gf\" (UniqueName: \"kubernetes.io/projected/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-kube-api-access-fv8gf\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621724 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-oauth-config\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.621770 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7406036-f795-40ba-b663-6b17adeabc8a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.622156 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-config\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.622270 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.622911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.622965 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6p4v\" (UniqueName: \"kubernetes.io/projected/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-kube-api-access-z6p4v\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.622995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.623300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-dir\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.623725 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.623780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kb6s\" (UniqueName: \"kubernetes.io/projected/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-kube-api-access-7kb6s\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.623926 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.623929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a80826b-26aa-45d4-991f-d1b585ab1a49-config\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.623975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.624012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl4z9\" (UniqueName: \"kubernetes.io/projected/677864f6-90af-4ffa-9b36-729194cedcd5-kube-api-access-jl4z9\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.624064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1de639b4-f614-41b9-812b-a7bfea3ccc90-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.624094 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.624708 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a80826b-26aa-45d4-991f-d1b585ab1a49-config\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625082 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-trusted-ca-bundle\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625223 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9371528e-0a6d-4946-bb4f-e4781e098460-config\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-policies\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ntw8\" (UniqueName: \"kubernetes.io/projected/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-kube-api-access-2ntw8\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625450 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8f413f0a-b891-46aa-975f-388300b1931d-metrics-tls\") pod \"dns-operator-744455d44c-shlh8\" (UID: \"8f413f0a-b891-46aa-975f-388300b1931d\") " pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89rhj\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-kube-api-access-89rhj\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625614 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-client-ca\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625707 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqdx2\" (UniqueName: \"kubernetes.io/projected/8f413f0a-b891-46aa-975f-388300b1931d-kube-api-access-dqdx2\") pod \"dns-operator-744455d44c-shlh8\" (UID: \"8f413f0a-b891-46aa-975f-388300b1931d\") " pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625740 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-client\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625792 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-service-ca\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.625963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-policies\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.626846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7406036-f795-40ba-b663-6b17adeabc8a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.626968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a80826b-26aa-45d4-991f-d1b585ab1a49-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627152 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627191 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-certificates\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627241 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmhtr\" (UniqueName: \"kubernetes.io/projected/c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4-kube-api-access-rmhtr\") pod \"downloads-7954f5f757-6fvxb\" (UID: \"c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4\") " pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-serving-cert\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwcms\" (UniqueName: \"kubernetes.io/projected/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-kube-api-access-mwcms\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627307 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627326 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/86de2a01-5e19-41e5-ba51-6acd28109d9c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-service-ca-bundle\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627380 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7406036-f795-40ba-b663-6b17adeabc8a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/86de2a01-5e19-41e5-ba51-6acd28109d9c-images\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-config-volume\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-metrics-certs\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8pmc\" (UniqueName: \"kubernetes.io/projected/b7406036-f795-40ba-b663-6b17adeabc8a-kube-api-access-b8pmc\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627462 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-oauth-serving-cert\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9371528e-0a6d-4946-bb4f-e4781e098460-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627496 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26pdg\" (UniqueName: \"kubernetes.io/projected/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-kube-api-access-26pdg\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627514 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-config\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627528 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9fzm\" (UniqueName: \"kubernetes.io/projected/41d11685-f01b-4304-a08e-3c43f21e1411-kube-api-access-d9fzm\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627558 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627573 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-console-config\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.627591 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677864f6-90af-4ffa-9b36-729194cedcd5-serving-cert\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: E1205 11:10:52.629425 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.129383633 +0000 UTC m=+148.520360191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.629709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.630146 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-oauth-serving-cert\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.630318 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-config\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.630427 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a80826b-26aa-45d4-991f-d1b585ab1a49-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.630446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-serving-cert\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.631346 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-console-config\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.631525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/86de2a01-5e19-41e5-ba51-6acd28109d9c-images\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.631562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-trusted-ca\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.632081 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/86de2a01-5e19-41e5-ba51-6acd28109d9c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.633114 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-client-ca\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.633675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9371528e-0a6d-4946-bb4f-e4781e098460-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.633735 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-metrics-tls\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.633870 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-serving-cert\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.633916 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.633968 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677864f6-90af-4ffa-9b36-729194cedcd5-serving-cert\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.633993 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8f413f0a-b891-46aa-975f-388300b1931d-metrics-tls\") pod \"dns-operator-744455d44c-shlh8\" (UID: \"8f413f0a-b891-46aa-975f-388300b1931d\") " pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634256 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/86de2a01-5e19-41e5-ba51-6acd28109d9c-proxy-tls\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634445 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-serving-cert\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-oauth-config\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b7406036-f795-40ba-b663-6b17adeabc8a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.634987 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.635424 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-config-volume\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.635844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.647580 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.647702 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.667615 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.678716 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.688425 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.690391 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.698943 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.706024 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.708858 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.715292 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.727538 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728007 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728215 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sszx7\" (UniqueName: \"kubernetes.io/projected/5e4c8c06-cfe8-4142-89f6-3a979c8135df-kube-api-access-sszx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7e9748a-1d74-4b39-9bf0-5c94dda32218-config-volume\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-stats-auth\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76d08b92-258a-4104-bca0-2ba684b8370d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728296 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fec6160-ac21-4265-b817-09954a91306a-serving-cert\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728311 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-srv-cert\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4cn4\" (UniqueName: \"kubernetes.io/projected/8781fe56-ed07-4998-b905-7fd3fc750e42-kube-api-access-l4cn4\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728342 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-webhook-cert\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728358 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-apiservice-cert\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728376 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9kbm\" (UniqueName: \"kubernetes.io/projected/45e13443-0823-411e-bac1-10eadf9b9918-kube-api-access-v9kbm\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728403 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30222533-a80a-4809-8cfb-971d5b8c4ac8-cert\") pod \"ingress-canary-8d57d\" (UID: \"30222533-a80a-4809-8cfb-971d5b8c4ac8\") " pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9864b4d8-c2ab-44ca-8d2b-cca19b047185-metrics-tls\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25567\" (UniqueName: \"kubernetes.io/projected/ddd951b5-5bbf-4f65-b09a-4482bb58fcb6-kube-api-access-25567\") pod \"package-server-manager-789f6589d5-7n7zc\" (UID: \"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7e9748a-1d74-4b39-9bf0-5c94dda32218-secret-volume\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728524 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-tmpfs\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728547 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1de639b4-f614-41b9-812b-a7bfea3ccc90-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728590 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-registration-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728605 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q648\" (UniqueName: \"kubernetes.io/projected/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-kube-api-access-5q648\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728675 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp6s4\" (UniqueName: \"kubernetes.io/projected/30222533-a80a-4809-8cfb-971d5b8c4ac8-kube-api-access-pp6s4\") pod \"ingress-canary-8d57d\" (UID: \"30222533-a80a-4809-8cfb-971d5b8c4ac8\") " pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89rhj\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-kube-api-access-89rhj\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b1f28ed0-5326-48e2-ab27-0caff604a251-certs\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-client\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728768 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/520f140b-3326-4ec2-8a13-a1142581c831-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c166a15f-8b67-4236-8fd2-f3e84449c404-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ml6sr\" (UID: \"c166a15f-8b67-4236-8fd2-f3e84449c404\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728822 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-certificates\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728934 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9864b4d8-c2ab-44ca-8d2b-cca19b047185-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.728985 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69dxb\" (UniqueName: \"kubernetes.io/projected/e7e9748a-1d74-4b39-9bf0-5c94dda32218-kube-api-access-69dxb\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729012 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndbs9\" (UniqueName: \"kubernetes.io/projected/0542946a-f040-447f-9461-9ea98f43a1d6-kube-api-access-ndbs9\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz2mp\" (UniqueName: \"kubernetes.io/projected/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-kube-api-access-kz2mp\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729066 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwcms\" (UniqueName: \"kubernetes.io/projected/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-kube-api-access-mwcms\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96d6f\" (UniqueName: \"kubernetes.io/projected/520f140b-3326-4ec2-8a13-a1142581c831-kube-api-access-96d6f\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729118 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b1f28ed0-5326-48e2-ab27-0caff604a251-node-bootstrap-token\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-service-ca-bundle\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e4c8c06-cfe8-4142-89f6-3a979c8135df-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729234 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-csi-data-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-metrics-certs\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxksf\" (UniqueName: \"kubernetes.io/projected/b1f28ed0-5326-48e2-ab27-0caff604a251-kube-api-access-sxksf\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26pdg\" (UniqueName: \"kubernetes.io/projected/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-kube-api-access-26pdg\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0542946a-f040-447f-9461-9ea98f43a1d6-signing-cabundle\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729394 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76d08b92-258a-4104-bca0-2ba684b8370d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729420 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729452 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/520f140b-3326-4ec2-8a13-a1142581c831-proxy-tls\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-tls\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729823 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6gtg\" (UniqueName: \"kubernetes.io/projected/7fec6160-ac21-4265-b817-09954a91306a-kube-api-access-c6gtg\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729884 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9864b4d8-c2ab-44ca-8d2b-cca19b047185-trusted-ca\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-bound-sa-token\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-544bb\" (UniqueName: \"kubernetes.io/projected/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-kube-api-access-544bb\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.729990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76d08b92-258a-4104-bca0-2ba684b8370d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.730018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/45e13443-0823-411e-bac1-10eadf9b9918-profile-collector-cert\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.730376 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1de639b4-f614-41b9-812b-a7bfea3ccc90-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.730517 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-certificates\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.731577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-default-certificate\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.731780 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:52 crc kubenswrapper[4809]: E1205 11:10:52.731899 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.231875646 +0000 UTC m=+148.622852264 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-config\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-service-ca-bundle\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732228 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-ca\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732285 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-socket-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732310 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-plugins-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732332 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/45e13443-0823-411e-bac1-10eadf9b9918-srv-cert\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0542946a-f040-447f-9461-9ea98f43a1d6-signing-key\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4zjv\" (UniqueName: \"kubernetes.io/projected/c166a15f-8b67-4236-8fd2-f3e84449c404-kube-api-access-t4zjv\") pod \"control-plane-machine-set-operator-78cbb6b69f-ml6sr\" (UID: \"c166a15f-8b67-4236-8fd2-f3e84449c404\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732465 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-mountpoint-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732493 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-serving-cert\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732530 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds829\" (UniqueName: \"kubernetes.io/projected/2c80d5ba-231a-4567-874c-116b797c940e-kube-api-access-ds829\") pod \"migrator-59844c95c7-rxf5n\" (UID: \"2c80d5ba-231a-4567-874c-116b797c940e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732582 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9mgh\" (UniqueName: \"kubernetes.io/projected/f8904696-72b4-4d00-83f1-552b9909afcf-kube-api-access-m9mgh\") pod \"multus-admission-controller-857f4d67dd-pbsgg\" (UID: \"f8904696-72b4-4d00-83f1-552b9909afcf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732609 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e4c8c06-cfe8-4142-89f6-3a979c8135df-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8s9d\" (UniqueName: \"kubernetes.io/projected/9864b4d8-c2ab-44ca-8d2b-cca19b047185-kube-api-access-d8s9d\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fec6160-ac21-4265-b817-09954a91306a-config\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8904696-72b4-4d00-83f1-552b9909afcf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pbsgg\" (UID: \"f8904696-72b4-4d00-83f1-552b9909afcf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732780 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd951b5-5bbf-4f65-b09a-4482bb58fcb6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7n7zc\" (UID: \"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732806 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-trusted-ca\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.732855 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1de639b4-f614-41b9-812b-a7bfea3ccc90-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.733375 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-service-ca\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.733449 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-ca\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.735980 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-config\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.736518 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-service-ca\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.737062 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-trusted-ca\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.742503 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1de639b4-f614-41b9-812b-a7bfea3ccc90-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.743299 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-tls\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.743695 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-metrics-certs\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.745300 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-stats-auth\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.747141 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-serving-cert\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.751805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-default-certificate\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.752192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-etcd-client\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.757509 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.768584 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.778676 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.779060 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.788557 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.801820 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.809317 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.810000 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.829974 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.857976 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.870261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7e9748a-1d74-4b39-9bf0-5c94dda32218-secret-volume\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871363 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871406 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-tmpfs\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-registration-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q648\" (UniqueName: \"kubernetes.io/projected/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-kube-api-access-5q648\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp6s4\" (UniqueName: \"kubernetes.io/projected/30222533-a80a-4809-8cfb-971d5b8c4ac8-kube-api-access-pp6s4\") pod \"ingress-canary-8d57d\" (UID: \"30222533-a80a-4809-8cfb-971d5b8c4ac8\") " pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/520f140b-3326-4ec2-8a13-a1142581c831-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b1f28ed0-5326-48e2-ab27-0caff604a251-certs\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c166a15f-8b67-4236-8fd2-f3e84449c404-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ml6sr\" (UID: \"c166a15f-8b67-4236-8fd2-f3e84449c404\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871744 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9864b4d8-c2ab-44ca-8d2b-cca19b047185-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871761 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69dxb\" (UniqueName: \"kubernetes.io/projected/e7e9748a-1d74-4b39-9bf0-5c94dda32218-kube-api-access-69dxb\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndbs9\" (UniqueName: \"kubernetes.io/projected/0542946a-f040-447f-9461-9ea98f43a1d6-kube-api-access-ndbs9\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz2mp\" (UniqueName: \"kubernetes.io/projected/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-kube-api-access-kz2mp\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871853 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96d6f\" (UniqueName: \"kubernetes.io/projected/520f140b-3326-4ec2-8a13-a1142581c831-kube-api-access-96d6f\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b1f28ed0-5326-48e2-ab27-0caff604a251-node-bootstrap-token\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871891 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871935 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e4c8c06-cfe8-4142-89f6-3a979c8135df-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871951 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-csi-data-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.871974 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxksf\" (UniqueName: \"kubernetes.io/projected/b1f28ed0-5326-48e2-ab27-0caff604a251-kube-api-access-sxksf\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0542946a-f040-447f-9461-9ea98f43a1d6-signing-cabundle\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872041 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76d08b92-258a-4104-bca0-2ba684b8370d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/520f140b-3326-4ec2-8a13-a1142581c831-proxy-tls\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6gtg\" (UniqueName: \"kubernetes.io/projected/7fec6160-ac21-4265-b817-09954a91306a-kube-api-access-c6gtg\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872178 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9864b4d8-c2ab-44ca-8d2b-cca19b047185-trusted-ca\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-544bb\" (UniqueName: \"kubernetes.io/projected/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-kube-api-access-544bb\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76d08b92-258a-4104-bca0-2ba684b8370d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/45e13443-0823-411e-bac1-10eadf9b9918-profile-collector-cert\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-socket-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-plugins-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872913 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0542946a-f040-447f-9461-9ea98f43a1d6-signing-key\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/45e13443-0823-411e-bac1-10eadf9b9918-srv-cert\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872949 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4zjv\" (UniqueName: \"kubernetes.io/projected/c166a15f-8b67-4236-8fd2-f3e84449c404-kube-api-access-t4zjv\") pod \"control-plane-machine-set-operator-78cbb6b69f-ml6sr\" (UID: \"c166a15f-8b67-4236-8fd2-f3e84449c404\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.872986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-mountpoint-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds829\" (UniqueName: \"kubernetes.io/projected/2c80d5ba-231a-4567-874c-116b797c940e-kube-api-access-ds829\") pod \"migrator-59844c95c7-rxf5n\" (UID: \"2c80d5ba-231a-4567-874c-116b797c940e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873023 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9mgh\" (UniqueName: \"kubernetes.io/projected/f8904696-72b4-4d00-83f1-552b9909afcf-kube-api-access-m9mgh\") pod \"multus-admission-controller-857f4d67dd-pbsgg\" (UID: \"f8904696-72b4-4d00-83f1-552b9909afcf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873108 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e4c8c06-cfe8-4142-89f6-3a979c8135df-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8s9d\" (UniqueName: \"kubernetes.io/projected/9864b4d8-c2ab-44ca-8d2b-cca19b047185-kube-api-access-d8s9d\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fec6160-ac21-4265-b817-09954a91306a-config\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873179 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8904696-72b4-4d00-83f1-552b9909afcf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pbsgg\" (UID: \"f8904696-72b4-4d00-83f1-552b9909afcf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd951b5-5bbf-4f65-b09a-4482bb58fcb6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7n7zc\" (UID: \"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sszx7\" (UniqueName: \"kubernetes.io/projected/5e4c8c06-cfe8-4142-89f6-3a979c8135df-kube-api-access-sszx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873341 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7e9748a-1d74-4b39-9bf0-5c94dda32218-config-volume\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fec6160-ac21-4265-b817-09954a91306a-serving-cert\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873419 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-srv-cert\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873442 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76d08b92-258a-4104-bca0-2ba684b8370d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4cn4\" (UniqueName: \"kubernetes.io/projected/8781fe56-ed07-4998-b905-7fd3fc750e42-kube-api-access-l4cn4\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873518 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-webhook-cert\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9kbm\" (UniqueName: \"kubernetes.io/projected/45e13443-0823-411e-bac1-10eadf9b9918-kube-api-access-v9kbm\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873584 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-apiservice-cert\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9864b4d8-c2ab-44ca-8d2b-cca19b047185-metrics-tls\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30222533-a80a-4809-8cfb-971d5b8c4ac8-cert\") pod \"ingress-canary-8d57d\" (UID: \"30222533-a80a-4809-8cfb-971d5b8c4ac8\") " pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.873756 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25567\" (UniqueName: \"kubernetes.io/projected/ddd951b5-5bbf-4f65-b09a-4482bb58fcb6-kube-api-access-25567\") pod \"package-server-manager-789f6589d5-7n7zc\" (UID: \"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.875298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fec6160-ac21-4265-b817-09954a91306a-config\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.877096 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7e9748a-1d74-4b39-9bf0-5c94dda32218-config-volume\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: E1205 11:10:52.878237 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.378220816 +0000 UTC m=+148.769197374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.880733 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-tmpfs\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.880864 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-plugins-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.895198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-socket-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.901779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/520f140b-3326-4ec2-8a13-a1142581c831-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.904213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fec6160-ac21-4265-b817-09954a91306a-serving-cert\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.904799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.904937 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7e9748a-1d74-4b39-9bf0-5c94dda32218-secret-volume\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.905257 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-mountpoint-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.905269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-registration-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.905320 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-srv-cert\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.905366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-csi-data-dir\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.905405 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.906112 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.908523 4809 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.915398 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-apiservice-cert\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.916429 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/45e13443-0823-411e-bac1-10eadf9b9918-profile-collector-cert\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.918324 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/520f140b-3326-4ec2-8a13-a1142581c831-proxy-tls\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.921475 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30222533-a80a-4809-8cfb-971d5b8c4ac8-cert\") pod \"ingress-canary-8d57d\" (UID: \"30222533-a80a-4809-8cfb-971d5b8c4ac8\") " pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.922297 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-webhook-cert\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.934445 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kpt8q"] Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.939752 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.951802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8904696-72b4-4d00-83f1-552b9909afcf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pbsgg\" (UID: \"f8904696-72b4-4d00-83f1-552b9909afcf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.952246 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.975547 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.976408 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 11:10:52 crc kubenswrapper[4809]: E1205 11:10:52.976714 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.476693475 +0000 UTC m=+148.867670033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:52 crc kubenswrapper[4809]: I1205 11:10:52.986988 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.015970 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.023453 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.031591 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.036606 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.046850 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.068699 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.072899 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm"] Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.084556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.085434 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.585417473 +0000 UTC m=+148.976394031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.087464 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.108159 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.119419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0542946a-f040-447f-9461-9ea98f43a1d6-signing-key\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.127231 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.136187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0542946a-f040-447f-9461-9ea98f43a1d6-signing-cabundle\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.146575 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.158330 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6vmbb"] Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.166796 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.173441 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b1f28ed0-5326-48e2-ab27-0caff604a251-certs\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.185580 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.185856 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.685828207 +0000 UTC m=+149.076804785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.186281 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.186836 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.686822285 +0000 UTC m=+149.077798843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.187096 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.197195 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f8b2x"] Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.207946 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.225403 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4"] Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.227872 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.239256 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b1f28ed0-5326-48e2-ab27-0caff604a251-node-bootstrap-token\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.247414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.267153 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.275962 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e4c8c06-cfe8-4142-89f6-3a979c8135df-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:53 crc kubenswrapper[4809]: W1205 11:10:53.284995 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-ce5cfe20704b71aff3c14e5877ffb641068f8a31ffce82d9a4dfcf93b00349eb WatchSource:0}: Error finding container ce5cfe20704b71aff3c14e5877ffb641068f8a31ffce82d9a4dfcf93b00349eb: Status 404 returned error can't find the container with id ce5cfe20704b71aff3c14e5877ffb641068f8a31ffce82d9a4dfcf93b00349eb Dec 05 11:10:53 crc kubenswrapper[4809]: W1205 11:10:53.286604 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73360c2b_d090_42c2_8800_94b0700ae627.slice/crio-d39af90d2cc1153e104d23b68c9e99bdd5b4ce95a0eb24b0c9be1892ecf14037 WatchSource:0}: Error finding container d39af90d2cc1153e104d23b68c9e99bdd5b4ce95a0eb24b0c9be1892ecf14037: Status 404 returned error can't find the container with id d39af90d2cc1153e104d23b68c9e99bdd5b4ce95a0eb24b0c9be1892ecf14037 Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.286903 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.287459 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.787445344 +0000 UTC m=+149.178421892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.287841 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: W1205 11:10:53.289997 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-a3ebd2708a38339e3b98e55312ba0e71a61788577409fc47e283f68be9f881de WatchSource:0}: Error finding container a3ebd2708a38339e3b98e55312ba0e71a61788577409fc47e283f68be9f881de: Status 404 returned error can't find the container with id a3ebd2708a38339e3b98e55312ba0e71a61788577409fc47e283f68be9f881de Dec 05 11:10:53 crc kubenswrapper[4809]: W1205 11:10:53.290876 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-2132bb00044c9d11150ef80919f1fb63f390998feff4257ace024ead866d77b4 WatchSource:0}: Error finding container 2132bb00044c9d11150ef80919f1fb63f390998feff4257ace024ead866d77b4: Status 404 returned error can't find the container with id 2132bb00044c9d11150ef80919f1fb63f390998feff4257ace024ead866d77b4 Dec 05 11:10:53 crc kubenswrapper[4809]: W1205 11:10:53.292328 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27d1916f_dfbe_4e36_963c_958803f4eb66.slice/crio-84df0d99a66eba422a95295de402d1ba29a1f9167561ad4baa194111c08ebbba WatchSource:0}: Error finding container 84df0d99a66eba422a95295de402d1ba29a1f9167561ad4baa194111c08ebbba: Status 404 returned error can't find the container with id 84df0d99a66eba422a95295de402d1ba29a1f9167561ad4baa194111c08ebbba Dec 05 11:10:53 crc kubenswrapper[4809]: W1205 11:10:53.293180 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c896868_d020_4367_aad2_9497223c6026.slice/crio-e2b3bb489cf56508d54f652dd9522dea088511216e04b6bf7571815de178cc47 WatchSource:0}: Error finding container e2b3bb489cf56508d54f652dd9522dea088511216e04b6bf7571815de178cc47: Status 404 returned error can't find the container with id e2b3bb489cf56508d54f652dd9522dea088511216e04b6bf7571815de178cc47 Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.308392 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.327221 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.339475 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e4c8c06-cfe8-4142-89f6-3a979c8135df-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.346901 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.367529 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.379740 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76d08b92-258a-4104-bca0-2ba684b8370d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.387510 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.388545 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.388992 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.88897935 +0000 UTC m=+149.279955908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.390689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76d08b92-258a-4104-bca0-2ba684b8370d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.407221 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.424932 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ce5cfe20704b71aff3c14e5877ffb641068f8a31ffce82d9a4dfcf93b00349eb"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.426517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2132bb00044c9d11150ef80919f1fb63f390998feff4257ace024ead866d77b4"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.426991 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.430514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" event={"ID":"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c","Type":"ContainerStarted","Data":"0968fac069d2dba587d069fab43d9386581c7fbe51cfb81c057aafe52288eb8a"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.434024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" event={"ID":"3c896868-d020-4367-aad2-9497223c6026","Type":"ContainerStarted","Data":"e2b3bb489cf56508d54f652dd9522dea088511216e04b6bf7571815de178cc47"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.437234 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gvwf"] Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.441678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c166a15f-8b67-4236-8fd2-f3e84449c404-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ml6sr\" (UID: \"c166a15f-8b67-4236-8fd2-f3e84449c404\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.451710 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.452389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a3ebd2708a38339e3b98e55312ba0e71a61788577409fc47e283f68be9f881de"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.473877 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" event={"ID":"2ddee1c1-dd55-429b-b277-214aa6556913","Type":"ContainerStarted","Data":"23d01dee58c9fe8a1b0ee12f8aa293cbfe08b4ba39c2aa3aea93d540b638e849"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.473941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" event={"ID":"2ddee1c1-dd55-429b-b277-214aa6556913","Type":"ContainerStarted","Data":"53d3fc9feeca6cb9bc42e7453a91f05689d4f2d085b569939b30de26e7082abf"} Dec 05 11:10:53 crc kubenswrapper[4809]: W1205 11:10:53.475273 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb52152a3_7674_489f_b67b_e96346e88228.slice/crio-97e8c535f2af5be6701590b3b314ec1611aafa61d82314e650bfb641c03e751c WatchSource:0}: Error finding container 97e8c535f2af5be6701590b3b314ec1611aafa61d82314e650bfb641c03e751c: Status 404 returned error can't find the container with id 97e8c535f2af5be6701590b3b314ec1611aafa61d82314e650bfb641c03e751c Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.476256 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.482755 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" event={"ID":"fceeabf7-b542-4506-a6c2-dae29b3ad4a3","Type":"ContainerStarted","Data":"6ef89521f90d0431b6f9ad117e3e02d7cfc81599d9349d153cbfbec20864718e"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.483034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" event={"ID":"fceeabf7-b542-4506-a6c2-dae29b3ad4a3","Type":"ContainerStarted","Data":"567a0f6cdc0903bca3cebc85be6645d44b1b2ce5ce7581fc6d4346692c55eb29"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.483343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/45e13443-0823-411e-bac1-10eadf9b9918-srv-cert\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.484801 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" event={"ID":"27d1916f-dfbe-4e36-963c-958803f4eb66","Type":"ContainerStarted","Data":"84df0d99a66eba422a95295de402d1ba29a1f9167561ad4baa194111c08ebbba"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.487647 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.489409 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.489643 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.98960698 +0000 UTC m=+149.380583538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.489821 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.490129 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:53.990116634 +0000 UTC m=+149.381093192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.491659 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" event={"ID":"73360c2b-d090-42c2-8800-94b0700ae627","Type":"ContainerStarted","Data":"d39af90d2cc1153e104d23b68c9e99bdd5b4ce95a0eb24b0c9be1892ecf14037"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.493838 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" event={"ID":"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978","Type":"ContainerStarted","Data":"01aa20d2121e26c9e3f6c2b170f86c88bf1e3004aaaa994be87bee01027bf31e"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.498994 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" event={"ID":"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2","Type":"ContainerStarted","Data":"252282f76619c82c3f2db95c2eb7ddfbfcf52b5b94999d400cb54ed7875b6f86"} Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.500514 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddd951b5-5bbf-4f65-b09a-4482bb58fcb6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7n7zc\" (UID: \"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.508214 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.527046 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.530797 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9864b4d8-c2ab-44ca-8d2b-cca19b047185-metrics-tls\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.552294 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.562243 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9864b4d8-c2ab-44ca-8d2b-cca19b047185-trusted-ca\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.569683 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.591572 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.591747 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.091721521 +0000 UTC m=+149.482698079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.592146 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.592493 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.092485993 +0000 UTC m=+149.483462551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.620551 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9371528e-0a6d-4946-bb4f-e4781e098460-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vt9cc\" (UID: \"9371528e-0a6d-4946-bb4f-e4781e098460\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.622705 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a80826b-26aa-45d4-991f-d1b585ab1a49-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lm6m8\" (UID: \"3a80826b-26aa-45d4-991f-d1b585ab1a49\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.625359 4809 request.go:700] Waited for 1.001447982s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-config-operator/serviceaccounts/openshift-config-operator/token Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.643378 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kb6s\" (UniqueName: \"kubernetes.io/projected/f0cdbeff-6bd6-4454-b0c4-1d556e43cf89-kube-api-access-7kb6s\") pod \"openshift-config-operator-7777fb866f-rwmwd\" (UID: \"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.662602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6p4v\" (UniqueName: \"kubernetes.io/projected/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-kube-api-access-z6p4v\") pod \"oauth-openshift-558db77b4-t4hqr\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.682549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl4z9\" (UniqueName: \"kubernetes.io/projected/677864f6-90af-4ffa-9b36-729194cedcd5-kube-api-access-jl4z9\") pod \"route-controller-manager-6576b87f9c-wlw76\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.693022 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.693411 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.19338336 +0000 UTC m=+149.584359958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.702602 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.703582 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqvgv\" (UniqueName: \"kubernetes.io/projected/86de2a01-5e19-41e5-ba51-6acd28109d9c-kube-api-access-zqvgv\") pod \"machine-config-operator-74547568cd-g2tb5\" (UID: \"86de2a01-5e19-41e5-ba51-6acd28109d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.723911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv8gf\" (UniqueName: \"kubernetes.io/projected/cd4e27b9-59fc-4024-b4ca-7e68a27605a9-kube-api-access-fv8gf\") pod \"dns-default-wml6l\" (UID: \"cd4e27b9-59fc-4024-b4ca-7e68a27605a9\") " pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.727921 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.741093 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ntw8\" (UniqueName: \"kubernetes.io/projected/0dab445a-a94d-4abb-8383-fe7b0ba3f2d9-kube-api-access-2ntw8\") pod \"console-operator-58897d9998-hhcsc\" (UID: \"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9\") " pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.763334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqdx2\" (UniqueName: \"kubernetes.io/projected/8f413f0a-b891-46aa-975f-388300b1931d-kube-api-access-dqdx2\") pod \"dns-operator-744455d44c-shlh8\" (UID: \"8f413f0a-b891-46aa-975f-388300b1931d\") " pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.782382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmhtr\" (UniqueName: \"kubernetes.io/projected/c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4-kube-api-access-rmhtr\") pod \"downloads-7954f5f757-6fvxb\" (UID: \"c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4\") " pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.800798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.801574 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.301558483 +0000 UTC m=+149.692535041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.803662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.806903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.816457 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.820128 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.821759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8pmc\" (UniqueName: \"kubernetes.io/projected/b7406036-f795-40ba-b663-6b17adeabc8a-kube-api-access-b8pmc\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.828210 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9fzm\" (UniqueName: \"kubernetes.io/projected/41d11685-f01b-4304-a08e-3c43f21e1411-kube-api-access-d9fzm\") pod \"console-f9d7485db-zhtzr\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.835089 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.844819 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7406036-f795-40ba-b663-6b17adeabc8a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hmv58\" (UID: \"b7406036-f795-40ba-b663-6b17adeabc8a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.861401 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwcms\" (UniqueName: \"kubernetes.io/projected/8d14e7df-f1f6-4a5d-964b-7f089e99d0fe-kube-api-access-mwcms\") pod \"router-default-5444994796-xgftq\" (UID: \"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe\") " pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.863826 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.871069 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.884269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-bound-sa-token\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.885955 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.902085 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:53 crc kubenswrapper[4809]: E1205 11:10:53.902554 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.402512622 +0000 UTC m=+149.793489180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.911382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26pdg\" (UniqueName: \"kubernetes.io/projected/47adcfd5-90bb-4792-bcb6-f0f01ac097f9-kube-api-access-26pdg\") pod \"etcd-operator-b45778765-qm6jb\" (UID: \"47adcfd5-90bb-4792-bcb6-f0f01ac097f9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.913407 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc"] Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.941824 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8"] Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.951727 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89rhj\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-kube-api-access-89rhj\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.968755 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:53 crc kubenswrapper[4809]: I1205 11:10:53.987039 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6gtg\" (UniqueName: \"kubernetes.io/projected/7fec6160-ac21-4265-b817-09954a91306a-kube-api-access-c6gtg\") pod \"service-ca-operator-777779d784-n4fnx\" (UID: \"7fec6160-ac21-4265-b817-09954a91306a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.003355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.003793 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.50377531 +0000 UTC m=+149.894751868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.011767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8s9d\" (UniqueName: \"kubernetes.io/projected/9864b4d8-c2ab-44ca-8d2b-cca19b047185-kube-api-access-d8s9d\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.028905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxksf\" (UniqueName: \"kubernetes.io/projected/b1f28ed0-5326-48e2-ab27-0caff604a251-kube-api-access-sxksf\") pod \"machine-config-server-cv89t\" (UID: \"b1f28ed0-5326-48e2-ab27-0caff604a251\") " pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.044319 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.047190 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp6s4\" (UniqueName: \"kubernetes.io/projected/30222533-a80a-4809-8cfb-971d5b8c4ac8-kube-api-access-pp6s4\") pod \"ingress-canary-8d57d\" (UID: \"30222533-a80a-4809-8cfb-971d5b8c4ac8\") " pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.064553 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.072240 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz2mp\" (UniqueName: \"kubernetes.io/projected/392dbe41-06c4-45cd-9e24-69ecc8f1b8e7-kube-api-access-kz2mp\") pod \"csi-hostpathplugin-2rpwq\" (UID: \"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7\") " pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.091197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.104828 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.105299 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.605278664 +0000 UTC m=+149.996255222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.107923 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds829\" (UniqueName: \"kubernetes.io/projected/2c80d5ba-231a-4567-874c-116b797c940e-kube-api-access-ds829\") pod \"migrator-59844c95c7-rxf5n\" (UID: \"2c80d5ba-231a-4567-874c-116b797c940e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.130181 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cv89t" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.146545 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.159216 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.159419 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.160523 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96d6f\" (UniqueName: \"kubernetes.io/projected/520f140b-3326-4ec2-8a13-a1142581c831-kube-api-access-96d6f\") pod \"machine-config-controller-84d6567774-pf4fm\" (UID: \"520f140b-3326-4ec2-8a13-a1142581c831\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.214423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.214740 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.214798 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.714783935 +0000 UTC m=+150.105760503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.225278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sszx7\" (UniqueName: \"kubernetes.io/projected/5e4c8c06-cfe8-4142-89f6-3a979c8135df-kube-api-access-sszx7\") pod \"kube-storage-version-migrator-operator-b67b599dd-bcsqv\" (UID: \"5e4c8c06-cfe8-4142-89f6-3a979c8135df\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.228957 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9mgh\" (UniqueName: \"kubernetes.io/projected/f8904696-72b4-4d00-83f1-552b9909afcf-kube-api-access-m9mgh\") pod \"multus-admission-controller-857f4d67dd-pbsgg\" (UID: \"f8904696-72b4-4d00-83f1-552b9909afcf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.230846 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hhcsc"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.234800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4zjv\" (UniqueName: \"kubernetes.io/projected/c166a15f-8b67-4236-8fd2-f3e84449c404-kube-api-access-t4zjv\") pod \"control-plane-machine-set-operator-78cbb6b69f-ml6sr\" (UID: \"c166a15f-8b67-4236-8fd2-f3e84449c404\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.236903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q648\" (UniqueName: \"kubernetes.io/projected/771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a-kube-api-access-5q648\") pod \"olm-operator-6b444d44fb-5rqh4\" (UID: \"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.240298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-544bb\" (UniqueName: \"kubernetes.io/projected/b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a-kube-api-access-544bb\") pod \"packageserver-d55dfcdfc-fhht6\" (UID: \"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.241467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69dxb\" (UniqueName: \"kubernetes.io/projected/e7e9748a-1d74-4b39-9bf0-5c94dda32218-kube-api-access-69dxb\") pod \"collect-profiles-29415540-ccjd8\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.247867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndbs9\" (UniqueName: \"kubernetes.io/projected/0542946a-f040-447f-9461-9ea98f43a1d6-kube-api-access-ndbs9\") pod \"service-ca-9c57cc56f-9b4bq\" (UID: \"0542946a-f040-447f-9461-9ea98f43a1d6\") " pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.248411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.313563 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25567\" (UniqueName: \"kubernetes.io/projected/ddd951b5-5bbf-4f65-b09a-4482bb58fcb6-kube-api-access-25567\") pod \"package-server-manager-789f6589d5-7n7zc\" (UID: \"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.315693 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.315787 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.815771495 +0000 UTC m=+150.206748053 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.316207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.316490 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.816483615 +0000 UTC m=+150.207460173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.315886 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8d57d" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.317118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9864b4d8-c2ab-44ca-8d2b-cca19b047185-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bstzg\" (UID: \"9864b4d8-c2ab-44ca-8d2b-cca19b047185\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.337377 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.344327 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4cn4\" (UniqueName: \"kubernetes.io/projected/8781fe56-ed07-4998-b905-7fd3fc750e42-kube-api-access-l4cn4\") pod \"marketplace-operator-79b997595-wfbxx\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.344398 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9kbm\" (UniqueName: \"kubernetes.io/projected/45e13443-0823-411e-bac1-10eadf9b9918-kube-api-access-v9kbm\") pod \"catalog-operator-68c6474976-l7dgh\" (UID: \"45e13443-0823-411e-bac1-10eadf9b9918\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.400439 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.412415 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.417489 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.417675 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.91764832 +0000 UTC m=+150.308624878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.417785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.418981 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:54.918965307 +0000 UTC m=+150.309941875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.419735 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.435479 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.451352 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.466089 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.466617 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.468137 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.478348 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.521167 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.521801 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.021784758 +0000 UTC m=+150.412761316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.530991 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wml6l"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.613608 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.623012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.623483 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.123468637 +0000 UTC m=+150.514445195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.714012 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-shlh8"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.724619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.724899 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.224884969 +0000 UTC m=+150.615861527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.724966 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.726319 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.226117524 +0000 UTC m=+150.617094082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.774799 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6fvxb"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.778871 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zhtzr"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.826863 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.827163 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.327150155 +0000 UTC m=+150.718126713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.827345 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.827680 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.327615418 +0000 UTC m=+150.718591976 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.980860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:54 crc kubenswrapper[4809]: E1205 11:10:54.981288 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.481269095 +0000 UTC m=+150.872245653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.991362 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.991905 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4hqr"] Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.995280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" event={"ID":"27d1916f-dfbe-4e36-963c-958803f4eb66","Type":"ContainerStarted","Data":"b1f996063cecf3560494a3b1f3d7b4c7ea760b26ffac449b4c1f8859f7fe3b8d"} Dec 05 11:10:54 crc kubenswrapper[4809]: I1205 11:10:54.996920 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd"] Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.034286 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76d08b92-258a-4104-bca0-2ba684b8370d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bvjsr\" (UID: \"76d08b92-258a-4104-bca0-2ba684b8370d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.046398 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.082622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" event={"ID":"3c896868-d020-4367-aad2-9497223c6026","Type":"ContainerStarted","Data":"c8ac973e77fdad83c5033e5a973bec5bf815c4a87db96c91e7014fa668e9c0e1"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.083893 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.084186 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.584173689 +0000 UTC m=+150.975150257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.169921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" event={"ID":"b52152a3-7674-489f-b67b-e96346e88228","Type":"ContainerStarted","Data":"65e97a7c38f9955063e988e7b6b2bb3f3c63535a0a25c0b54b0b304393df7495"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.169965 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" event={"ID":"b52152a3-7674-489f-b67b-e96346e88228","Type":"ContainerStarted","Data":"97e8c535f2af5be6701590b3b314ec1611aafa61d82314e650bfb641c03e751c"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.170265 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.173352 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" event={"ID":"2ddee1c1-dd55-429b-b277-214aa6556913","Type":"ContainerStarted","Data":"3c2f546600f0a1c02950354ffea5f80fb9a438fd38685f00764c29b797440347"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.177355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" event={"ID":"3a80826b-26aa-45d4-991f-d1b585ab1a49","Type":"ContainerStarted","Data":"6e42b4dc553e576c697d8e18c5f07dda09cf821edca22e67139743c8b88e806c"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.179304 4809 generic.go:334] "Generic (PLEG): container finished" podID="d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2" containerID="327f2369947b7f93cbb81b18f65659c2b4c7b68a2e251a9bbaada07e45994ee9" exitCode=0 Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.179386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" event={"ID":"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2","Type":"ContainerDied","Data":"327f2369947b7f93cbb81b18f65659c2b4c7b68a2e251a9bbaada07e45994ee9"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.181032 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" event={"ID":"9371528e-0a6d-4946-bb4f-e4781e098460","Type":"ContainerStarted","Data":"a3ec4684602a10734f448ac9bbb9ef5e5c3cfe113a9cf25ffbddb249cee4762f"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.182323 4809 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7gvwf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.182371 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" podUID="b52152a3-7674-489f-b67b-e96346e88228" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.186586 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.187447 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.687424443 +0000 UTC m=+151.078401051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.191290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"10250566b5a8ca1dbd1e06038e4f8b027be601a9a75eaac6b1bd92237741c04c"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.216913 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ce9f9e958ea89dab2f4c5e05f423a11f7841ec4b49047e301bfd0745e35c28ff"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.217419 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:10:55 crc kubenswrapper[4809]: W1205 11:10:55.217616 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod677864f6_90af_4ffa_9b36_729194cedcd5.slice/crio-0ab14695c81fc30b85012971dd3086c04ea7bb42296805c1ed19d66a496ed413 WatchSource:0}: Error finding container 0ab14695c81fc30b85012971dd3086c04ea7bb42296805c1ed19d66a496ed413: Status 404 returned error can't find the container with id 0ab14695c81fc30b85012971dd3086c04ea7bb42296805c1ed19d66a496ed413 Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.219353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" event={"ID":"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c","Type":"ContainerStarted","Data":"29c1ce2667e13489b728d941d65fd7a993113b9bc615717493523d2bcda193c9"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.219389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" event={"ID":"9cf0c441-26db-4ac8-a1ed-4c2c90bc214c","Type":"ContainerStarted","Data":"60a17daa925bdab8d92c582ece7f48a7c8374c119471d17ccfa4e0557ca0d980"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.239027 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e7bba4ec724cd12d9b4a5140f7594941f9fc3a4351034411f443c01d875e6dbb"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.243190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" event={"ID":"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978","Type":"ContainerStarted","Data":"5b0d74227c31e1f11a1819c20d989e3722c362d9ca15035123cae7a76170ab4d"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.287755 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.290964 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.790948904 +0000 UTC m=+151.181925462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.313187 4809 generic.go:334] "Generic (PLEG): container finished" podID="73360c2b-d090-42c2-8800-94b0700ae627" containerID="ce8eea09eb7e5423c7206f59163b3f068b87f6d40477047581a080989f8649a0" exitCode=0 Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.314382 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" event={"ID":"73360c2b-d090-42c2-8800-94b0700ae627","Type":"ContainerDied","Data":"ce8eea09eb7e5423c7206f59163b3f068b87f6d40477047581a080989f8649a0"} Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.391126 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.391459 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.89144507 +0000 UTC m=+151.282421628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.492818 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.494276 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:55.994259742 +0000 UTC m=+151.385236300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.515755 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx"] Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.567446 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qm6jb"] Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.594124 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.594399 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.094383447 +0000 UTC m=+151.485360005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.594435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.594697 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.094691386 +0000 UTC m=+151.485667944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: W1205 11:10:55.658614 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fec6160_ac21_4265_b817_09954a91306a.slice/crio-41f3f77acea655cb47ed113374d7d9c24183159852a300e4876c5aec49d77f13 WatchSource:0}: Error finding container 41f3f77acea655cb47ed113374d7d9c24183159852a300e4876c5aec49d77f13: Status 404 returned error can't find the container with id 41f3f77acea655cb47ed113374d7d9c24183159852a300e4876c5aec49d77f13 Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.720200 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.720913 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.220892708 +0000 UTC m=+151.611869266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.761293 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7l888" podStartSLOduration=127.761271847 podStartE2EDuration="2m7.761271847s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:55.760570767 +0000 UTC m=+151.151547325" watchObservedRunningTime="2025-12-05 11:10:55.761271847 +0000 UTC m=+151.152248425" Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.822085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.822468 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.322452914 +0000 UTC m=+151.713429472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:55 crc kubenswrapper[4809]: I1205 11:10:55.925334 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:55 crc kubenswrapper[4809]: E1205 11:10:55.925729 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.425690077 +0000 UTC m=+151.816666665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.026871 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.027477 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.527465909 +0000 UTC m=+151.918442467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.044924 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" podStartSLOduration=128.044901581 podStartE2EDuration="2m8.044901581s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:56.043194333 +0000 UTC m=+151.434170891" watchObservedRunningTime="2025-12-05 11:10:56.044901581 +0000 UTC m=+151.435878169" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.084020 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-f8b2x" podStartSLOduration=128.084005185 podStartE2EDuration="2m8.084005185s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:56.083070138 +0000 UTC m=+151.474046696" watchObservedRunningTime="2025-12-05 11:10:56.084005185 +0000 UTC m=+151.474981743" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.128168 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.128563 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.628544582 +0000 UTC m=+152.019521140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.160102 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8fcv4" podStartSLOduration=128.160080722 podStartE2EDuration="2m8.160080722s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:56.157153749 +0000 UTC m=+151.548130307" watchObservedRunningTime="2025-12-05 11:10:56.160080722 +0000 UTC m=+151.551057280" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.231377 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.231798 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.731782125 +0000 UTC m=+152.122758673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.334687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.335362 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.835344128 +0000 UTC m=+152.226320686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.401758 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hhcsc" event={"ID":"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9","Type":"ContainerStarted","Data":"c672a2cf2602791e3271b9f660155d768f296a2055dd5466f1a7d4ae7879ff7f"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.425963 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" event={"ID":"86de2a01-5e19-41e5-ba51-6acd28109d9c","Type":"ContainerStarted","Data":"3315b5181adc853927b65e8df70b537321296d0786596ab721a96ddafd1d65b8"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.426014 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" event={"ID":"86de2a01-5e19-41e5-ba51-6acd28109d9c","Type":"ContainerStarted","Data":"dfa015de9e620ce125a2f7bcee70c0821422c04ca495d15470cf04cbdd4e782d"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.426853 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gcwbn" podStartSLOduration=128.42684226 podStartE2EDuration="2m8.42684226s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:56.419554364 +0000 UTC m=+151.810530912" watchObservedRunningTime="2025-12-05 11:10:56.42684226 +0000 UTC m=+151.817818818" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.439601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.439978 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:56.93996448 +0000 UTC m=+152.330941038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.468215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" event={"ID":"677864f6-90af-4ffa-9b36-729194cedcd5","Type":"ContainerStarted","Data":"0ab14695c81fc30b85012971dd3086c04ea7bb42296805c1ed19d66a496ed413"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.505248 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cv89t" event={"ID":"b1f28ed0-5326-48e2-ab27-0caff604a251","Type":"ContainerStarted","Data":"42aa1fb20fe3ee247da0d51e648fe39e41a92758aaedd89e19509e5b595f9aab"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.520108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" event={"ID":"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b","Type":"ContainerStarted","Data":"d9201b743f1677e7e244b5af02f92edb08995cff5ec6b34dea455e6a9ec2e7a2"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.543162 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.543462 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.04343111 +0000 UTC m=+152.434407658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.570714 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm"] Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.577716 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wfbxx"] Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.627139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" event={"ID":"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89","Type":"ContainerStarted","Data":"a3ceacd9ae096c5f98bfb40547d7dbab288dcd6a21aba1bb0ecb42fc1eb614e7"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.648607 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.648917 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.148907287 +0000 UTC m=+152.539883845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.651872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zhtzr" event={"ID":"41d11685-f01b-4304-a08e-3c43f21e1411","Type":"ContainerStarted","Data":"42d7d9ea603b91cbdb34519c68964ce30eebc14f99152efe33aaa89895004962"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.656876 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6"] Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.677250 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" event={"ID":"8f413f0a-b891-46aa-975f-388300b1931d","Type":"ContainerStarted","Data":"ba66e2d96c515367fe58040b229980dd5bab591ce6a4de77a0375372bbf3f39e"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.678980 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6vmbb" podStartSLOduration=128.678962965 podStartE2EDuration="2m8.678962965s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:56.677305658 +0000 UTC m=+152.068282236" watchObservedRunningTime="2025-12-05 11:10:56.678962965 +0000 UTC m=+152.069939523" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.720114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xgftq" event={"ID":"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe","Type":"ContainerStarted","Data":"c8c0c3f001c0e17e6cf9a6634395aef5cdc1999b43a43f30044943a8f412d3d9"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.752279 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.761003 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.260978559 +0000 UTC m=+152.651955117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.766996 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" event={"ID":"47adcfd5-90bb-4792-bcb6-f0f01ac097f9","Type":"ContainerStarted","Data":"e03c582508fbd2571f4b808afc567e67289e8e89705928992f7466bc8e2a6bb7"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.773229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.774596 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.274583244 +0000 UTC m=+152.665559802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.796615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" event={"ID":"7fec6160-ac21-4265-b817-09954a91306a","Type":"ContainerStarted","Data":"41f3f77acea655cb47ed113374d7d9c24183159852a300e4876c5aec49d77f13"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.816786 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6fvxb" event={"ID":"c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4","Type":"ContainerStarted","Data":"8c164ff2007267d6fe6c279bbfd4b9476545e6dd05aec43ddc8286e47ec58730"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.834230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wml6l" event={"ID":"cd4e27b9-59fc-4024-b4ca-7e68a27605a9","Type":"ContainerStarted","Data":"7579d8c2c47c3a9b91adb6da2a152ede932b7f1c4f9027480cc14c3d32926975"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.840140 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-xgftq" podStartSLOduration=128.840112933 podStartE2EDuration="2m8.840112933s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:56.825203192 +0000 UTC m=+152.216179760" watchObservedRunningTime="2025-12-05 11:10:56.840112933 +0000 UTC m=+152.231089491" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.874770 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.875252 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.375238434 +0000 UTC m=+152.766214992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.926852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" event={"ID":"3a80826b-26aa-45d4-991f-d1b585ab1a49","Type":"ContainerStarted","Data":"e1f9095cb9a4c789efd41504138f90529afc1fc6c96dfc2e505110b437c0a648"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.940739 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" event={"ID":"9371528e-0a6d-4946-bb4f-e4781e098460","Type":"ContainerStarted","Data":"201f870d98c382cdd3b7249fcefe59e9b379bea28808522bd351752ad0264e3c"} Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.958725 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.970380 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lm6m8" podStartSLOduration=128.970350798 podStartE2EDuration="2m8.970350798s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:56.969294929 +0000 UTC m=+152.360271497" watchObservedRunningTime="2025-12-05 11:10:56.970350798 +0000 UTC m=+152.361327356" Dec 05 11:10:56 crc kubenswrapper[4809]: I1205 11:10:56.977415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:56 crc kubenswrapper[4809]: E1205 11:10:56.984921 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.484899899 +0000 UTC m=+152.875876457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.029810 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vt9cc" podStartSLOduration=129.029774845 podStartE2EDuration="2m9.029774845s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:57.023181409 +0000 UTC m=+152.414157967" watchObservedRunningTime="2025-12-05 11:10:57.029774845 +0000 UTC m=+152.420751423" Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.096080 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.097957 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.597940209 +0000 UTC m=+152.988916767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.135211 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.161031 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.168676 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.168753 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.238227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.238744 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.738728292 +0000 UTC m=+153.129704850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.261366 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.262462 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9b4bq"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.338919 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.339312 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.83928546 +0000 UTC m=+153.230262268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.341435 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.438751 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2rpwq"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.453214 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.458896 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.459528 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:57.959492592 +0000 UTC m=+153.350469150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.483094 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.570216 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.570968 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.070944948 +0000 UTC m=+153.461921516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.587773 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg"] Dec 05 11:10:57 crc kubenswrapper[4809]: W1205 11:10:57.613848 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7406036_f795_40ba_b663_6b17adeabc8a.slice/crio-730c60ba5f8448b7201c68494de8b53421e08b339ada98b854bdf9ddeacd9c8a WatchSource:0}: Error finding container 730c60ba5f8448b7201c68494de8b53421e08b339ada98b854bdf9ddeacd9c8a: Status 404 returned error can't find the container with id 730c60ba5f8448b7201c68494de8b53421e08b339ada98b854bdf9ddeacd9c8a Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.671794 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.672143 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.172130263 +0000 UTC m=+153.563106821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.725318 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8d57d"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.746227 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.772184 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.773997 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.774295 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.274280186 +0000 UTC m=+153.665256744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.793238 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.818949 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pbsgg"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.875599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.875896 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.375885484 +0000 UTC m=+153.766862042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.914377 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh"] Dec 05 11:10:57 crc kubenswrapper[4809]: I1205 11:10:57.977543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:57 crc kubenswrapper[4809]: E1205 11:10:57.977978 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.477955694 +0000 UTC m=+153.868932252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.078776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.079545 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.57952901 +0000 UTC m=+153.970505578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.082011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" event={"ID":"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b","Type":"ContainerStarted","Data":"8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.083213 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.099197 4809 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-t4hqr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.19:6443/healthz\": dial tcp 10.217.0.19:6443: connect: connection refused" start-of-body= Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.099261 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.19:6443/healthz\": dial tcp 10.217.0.19:6443: connect: connection refused" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.127111 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" podStartSLOduration=130.127094992 podStartE2EDuration="2m10.127094992s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.126295749 +0000 UTC m=+153.517272307" watchObservedRunningTime="2025-12-05 11:10:58.127094992 +0000 UTC m=+153.518071550" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.127389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" event={"ID":"73360c2b-d090-42c2-8800-94b0700ae627","Type":"ContainerStarted","Data":"c5934639f735097ff8a879e18c94d128d4af8b8bac8da8b6294a8d118b540a3b"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.139761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" event={"ID":"2c80d5ba-231a-4567-874c-116b797c940e","Type":"ContainerStarted","Data":"831829b950c274714256802471c84bcfc129a9eab1b6fd296872d791a20ee8f4"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.141818 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" event={"ID":"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a","Type":"ContainerStarted","Data":"b284d85035fd1c813037bd9c179554ec1341a7ba848fe6b64c6b817b5b4087f7"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.141864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" event={"ID":"b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a","Type":"ContainerStarted","Data":"053ab3a163e43fcba1689f3977b21de9457ed4a3eb8a94f91a1794a807d67a32"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.142859 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.144345 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" event={"ID":"520f140b-3326-4ec2-8a13-a1142581c831","Type":"ContainerStarted","Data":"f390cbb2d781d7e18a558d5c5fdcfa7c9d16f8c5126b3f2a3ebaeb341bb2c0cb"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.144371 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" event={"ID":"520f140b-3326-4ec2-8a13-a1142581c831","Type":"ContainerStarted","Data":"ea84e00b191e8b9c7a1615a2e0cd2b4787c88409b12677c61f5480c919fe84fc"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.149094 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" event={"ID":"0542946a-f040-447f-9461-9ea98f43a1d6","Type":"ContainerStarted","Data":"7ebba2a6eb6b731c5f554fb678b5e8b71e9224fc806f46bca8143d61799044e2"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.153991 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" event={"ID":"47adcfd5-90bb-4792-bcb6-f0f01ac097f9","Type":"ContainerStarted","Data":"76486d7909ce0e8273b5ab1059d93b0db1bbc38fcf90ceaef2d0f755855dd071"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.162606 4809 generic.go:334] "Generic (PLEG): container finished" podID="f0cdbeff-6bd6-4454-b0c4-1d556e43cf89" containerID="9e0dc6c9a7787ed8ff07af7731670c050cda0c4dd74b90e3a2bc92f9e7d551f5" exitCode=0 Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.162688 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" event={"ID":"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89","Type":"ContainerDied","Data":"9e0dc6c9a7787ed8ff07af7731670c050cda0c4dd74b90e3a2bc92f9e7d551f5"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.162702 4809 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-fhht6 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.162736 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" podUID="b7c6ff8b-fb59-47d6-99b9-c0d148a09a7a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.174783 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:10:58 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:10:58 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:10:58 crc kubenswrapper[4809]: healthz check failed Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.174830 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.180053 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.181134 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.681116617 +0000 UTC m=+154.072093175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.249039 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" event={"ID":"c166a15f-8b67-4236-8fd2-f3e84449c404","Type":"ContainerStarted","Data":"5484d24f89d158b618bf6f4fa3b027aefb6cf49268e27f4175ffb96ba51755ca"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.249099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" event={"ID":"c166a15f-8b67-4236-8fd2-f3e84449c404","Type":"ContainerStarted","Data":"c255befe27f6896d4d76327c50929220a4ec69e18ba047cfb00440e89f99b752"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.258932 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" podStartSLOduration=130.258917822 podStartE2EDuration="2m10.258917822s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.220076806 +0000 UTC m=+153.611053384" watchObservedRunningTime="2025-12-05 11:10:58.258917822 +0000 UTC m=+153.649894380" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.276968 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6fvxb" event={"ID":"c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4","Type":"ContainerStarted","Data":"5a5dcf754013110a47bba8ad28287d43e7b8422c6b930367495bfd8465f881eb"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.278221 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.282536 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.299777 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.799764305 +0000 UTC m=+154.190740863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.320270 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.320329 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.321896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" event={"ID":"5e4c8c06-cfe8-4142-89f6-3a979c8135df","Type":"ContainerStarted","Data":"7b81357eed7824472d3e55df574e04d61abe3b7fe2ad1c3f563f379de8535174"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.339591 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qm6jb" podStartSLOduration=130.339572938 podStartE2EDuration="2m10.339572938s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.338371454 +0000 UTC m=+153.729348012" watchObservedRunningTime="2025-12-05 11:10:58.339572938 +0000 UTC m=+153.730549496" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.355243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" event={"ID":"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2","Type":"ContainerStarted","Data":"05881a07a40b196c2e3cad2db6a42544b5a0e51a94517236846d4ec1e9e6e804"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.377282 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" event={"ID":"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7","Type":"ContainerStarted","Data":"e4359aa2f1be94d9b7dead8dcdbddf9f9d59af0db73f0c76c1fd8e4b42f958a7"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.386906 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.388234 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.888215291 +0000 UTC m=+154.279191849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.392595 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" podStartSLOduration=130.392582374 podStartE2EDuration="2m10.392582374s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.39102856 +0000 UTC m=+153.782005118" watchObservedRunningTime="2025-12-05 11:10:58.392582374 +0000 UTC m=+153.783558932" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.400158 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xgftq" event={"ID":"8d14e7df-f1f6-4a5d-964b-7f089e99d0fe","Type":"ContainerStarted","Data":"ba64f3d7cc8a70ae30fc5c0331f676f9f1c8afab0a3e9797ed2fb82eb5e190a8"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.413065 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" event={"ID":"8781fe56-ed07-4998-b905-7fd3fc750e42","Type":"ContainerStarted","Data":"1a746b32726ed8c572c8ecf84ccee922f17e05b980e33563e411bc7a0f460d7e"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.413119 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" event={"ID":"8781fe56-ed07-4998-b905-7fd3fc750e42","Type":"ContainerStarted","Data":"2da6558d17dcb8d098d6a5970114f9ffc3142551d1f3e3aac22bd29d6b99ce28"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.413795 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.422596 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ml6sr" podStartSLOduration=130.422579541 podStartE2EDuration="2m10.422579541s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.422251862 +0000 UTC m=+153.813228420" watchObservedRunningTime="2025-12-05 11:10:58.422579541 +0000 UTC m=+153.813556099" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.435476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" event={"ID":"677864f6-90af-4ffa-9b36-729194cedcd5","Type":"ContainerStarted","Data":"899a2699a3be2d750c5149359415a2fae24c1871ea328ddab5cdc861d437e58a"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.435934 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.452575 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wfbxx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.452626 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.463788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" event={"ID":"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a","Type":"ContainerStarted","Data":"c4c16c4a1905d4307aa437eb8afec0051c6044a70c0cdb1863229e83902ad486"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.464713 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.465542 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6fvxb" podStartSLOduration=130.465523313 podStartE2EDuration="2m10.465523313s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.458978938 +0000 UTC m=+153.849955496" watchObservedRunningTime="2025-12-05 11:10:58.465523313 +0000 UTC m=+153.856499871" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.488367 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5rqh4 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.488433 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" podUID="771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.489490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.491322 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:58.9913086 +0000 UTC m=+154.382285158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.493030 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" podStartSLOduration=130.490779656 podStartE2EDuration="2m10.490779656s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.488558713 +0000 UTC m=+153.879535271" watchObservedRunningTime="2025-12-05 11:10:58.490779656 +0000 UTC m=+153.881756214" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.497825 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" event={"ID":"7fec6160-ac21-4265-b817-09954a91306a","Type":"ContainerStarted","Data":"8b87d5abc20893b191a4d6901f23b0872ce49a64797bc78ddb7f641eafbfeaf9"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.522266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cv89t" event={"ID":"b1f28ed0-5326-48e2-ab27-0caff604a251","Type":"ContainerStarted","Data":"4b592bad6ae8beee4e535c0481c24848ba7ffa68adf7ccd0c67b18420ec07d9b"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.536043 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" podStartSLOduration=130.536028003 podStartE2EDuration="2m10.536028003s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.528315705 +0000 UTC m=+153.919292273" watchObservedRunningTime="2025-12-05 11:10:58.536028003 +0000 UTC m=+153.927004561" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.537754 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" event={"ID":"9864b4d8-c2ab-44ca-8d2b-cca19b047185","Type":"ContainerStarted","Data":"d7b14ba826af7287f27ad1728b993a5bf0648a116f8de55d04c05e3ca1449c60"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.551030 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" event={"ID":"e7e9748a-1d74-4b39-9bf0-5c94dda32218","Type":"ContainerStarted","Data":"034525a693eef5658a15b496f3a8691b5c169f3a632261af5bfd075fccfc87b3"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.553256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" event={"ID":"86de2a01-5e19-41e5-ba51-6acd28109d9c","Type":"ContainerStarted","Data":"10e13d8fb8d82edf06d82eba6fe6bef5b140899aab6b82e55d2e8907c125c29b"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.563150 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n4fnx" podStartSLOduration=130.563133327 podStartE2EDuration="2m10.563133327s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.562776007 +0000 UTC m=+153.953752605" watchObservedRunningTime="2025-12-05 11:10:58.563133327 +0000 UTC m=+153.954109885" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.581313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" event={"ID":"0d4a4cfe-eb08-4b55-8a2d-a87d2fd63978","Type":"ContainerStarted","Data":"c87c65336bb901fb5be122151dcc944e0981fe3a260a9159a47e1a934d628219"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.590259 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" podStartSLOduration=130.590236312 podStartE2EDuration="2m10.590236312s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.587172866 +0000 UTC m=+153.978149424" watchObservedRunningTime="2025-12-05 11:10:58.590236312 +0000 UTC m=+153.981212870" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.596150 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.597037 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.097016594 +0000 UTC m=+154.487993152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.624120 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.625735 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" event={"ID":"8f413f0a-b891-46aa-975f-388300b1931d","Type":"ContainerStarted","Data":"182a8e069f3224a8a4f178ee9f31c6b14d3ffc3c915d3cb5eaaa9db15df595ff"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.644439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zhtzr" event={"ID":"41d11685-f01b-4304-a08e-3c43f21e1411","Type":"ContainerStarted","Data":"83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.645268 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-cv89t" podStartSLOduration=7.645253975 podStartE2EDuration="7.645253975s" podCreationTimestamp="2025-12-05 11:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.644173134 +0000 UTC m=+154.035149702" watchObservedRunningTime="2025-12-05 11:10:58.645253975 +0000 UTC m=+154.036230523" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.656619 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" event={"ID":"b7406036-f795-40ba-b663-6b17adeabc8a","Type":"ContainerStarted","Data":"730c60ba5f8448b7201c68494de8b53421e08b339ada98b854bdf9ddeacd9c8a"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.669209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" event={"ID":"76d08b92-258a-4104-bca0-2ba684b8370d","Type":"ContainerStarted","Data":"da64792f1766b7f9984558d9400dfff8616256d39fd123c7ffcd7049a2411aee"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.732044 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wml6l" event={"ID":"cd4e27b9-59fc-4024-b4ca-7e68a27605a9","Type":"ContainerStarted","Data":"8db13e63788d898b3af9e15b1cc5eb215fbf9c1301a00df38ce24cb4bcd8dc2c"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.732741 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-df5wd" podStartSLOduration=130.732719073 podStartE2EDuration="2m10.732719073s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.729586835 +0000 UTC m=+154.120563393" watchObservedRunningTime="2025-12-05 11:10:58.732719073 +0000 UTC m=+154.123695631" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.733867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.735141 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.235123021 +0000 UTC m=+154.626099579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.747864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" event={"ID":"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6","Type":"ContainerStarted","Data":"f705f7639708f64b34ef7febaa62e09c9e2b06ba0be07cc9fb69d362c16e59ae"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.788499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hhcsc" event={"ID":"0dab445a-a94d-4abb-8383-fe7b0ba3f2d9","Type":"ContainerStarted","Data":"b3b0e7b432b99cc4fb31f64ba6cab53dc4554c01a0e95e60203dae7d6515a670"} Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.788542 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.839036 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-hhcsc" podStartSLOduration=130.839012863 podStartE2EDuration="2m10.839012863s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.835915686 +0000 UTC m=+154.226892264" watchObservedRunningTime="2025-12-05 11:10:58.839012863 +0000 UTC m=+154.229989431" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.841929 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g2tb5" podStartSLOduration=130.841916815 podStartE2EDuration="2m10.841916815s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.804220941 +0000 UTC m=+154.195197509" watchObservedRunningTime="2025-12-05 11:10:58.841916815 +0000 UTC m=+154.232893373" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.852001 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.852180 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.352124173 +0000 UTC m=+154.743100731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.852340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.853384 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.353370748 +0000 UTC m=+154.744347406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.889444 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zhtzr" podStartSLOduration=130.889407645 podStartE2EDuration="2m10.889407645s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:10:58.886114432 +0000 UTC m=+154.277090990" watchObservedRunningTime="2025-12-05 11:10:58.889407645 +0000 UTC m=+154.280384203" Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.953316 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.953442 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.453408701 +0000 UTC m=+154.844385259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:58 crc kubenswrapper[4809]: I1205 11:10:58.955738 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:58 crc kubenswrapper[4809]: E1205 11:10:58.961220 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.461202551 +0000 UTC m=+154.852179109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.056654 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.057215 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.55717817 +0000 UTC m=+154.948154728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.157905 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.158496 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.658485529 +0000 UTC m=+155.049462087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.222016 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:10:59 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:10:59 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:10:59 crc kubenswrapper[4809]: healthz check failed Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.222089 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.264889 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.265277 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.765258192 +0000 UTC m=+155.156234750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.337028 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-hhcsc" Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.366189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.366481 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.866469308 +0000 UTC m=+155.257445866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.468402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.468789 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.968762795 +0000 UTC m=+155.359739363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.469121 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.469579 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:10:59.969564938 +0000 UTC m=+155.360541496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.569938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.570552 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.070531727 +0000 UTC m=+155.461508285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.671656 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.671938 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.171919779 +0000 UTC m=+155.562896337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.775035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.775206 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.275183353 +0000 UTC m=+155.666159921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.775508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.775910 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.275894283 +0000 UTC m=+155.666870841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.830650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" event={"ID":"76d08b92-258a-4104-bca0-2ba684b8370d","Type":"ContainerStarted","Data":"0bf4c146f7254ba4efc33686910e4d3dda942d8a2b8c33f2359f87b763f5ad48"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.832917 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" event={"ID":"0542946a-f040-447f-9461-9ea98f43a1d6","Type":"ContainerStarted","Data":"ee916986c81cfec8ee08dcad8213c8ee4cfbceea608e80d65c8a280e5a38cfcc"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.834804 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wml6l" event={"ID":"cd4e27b9-59fc-4024-b4ca-7e68a27605a9","Type":"ContainerStarted","Data":"23141f4f2b1c661c736a625fc0cd120e4a3dfa3dfbee7fb5b305bec3aa121902"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.835266 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-wml6l" Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.836435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8d57d" event={"ID":"30222533-a80a-4809-8cfb-971d5b8c4ac8","Type":"ContainerStarted","Data":"f86861f1ab9ccaa1f1c15c7c72bd26d1954f5d60adfe24d1a3f679e80a62e7cb"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.836458 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8d57d" event={"ID":"30222533-a80a-4809-8cfb-971d5b8c4ac8","Type":"ContainerStarted","Data":"ee80673e34384d6f0ea274f0fe15b7512d598270a94850444734a312e784a6d7"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.837799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" event={"ID":"5e4c8c06-cfe8-4142-89f6-3a979c8135df","Type":"ContainerStarted","Data":"b9cbfc0f534994b9d27ea34a6a1c09c0241e26b272470565a2d07c6a0ca881c5"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.839418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" event={"ID":"771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a","Type":"ContainerStarted","Data":"5bb088e396e6da083eab1ec99e8066c9e942ab657c08e731699d273f65e05cb0"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.842945 4809 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5rqh4 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.842998 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" podUID="771e0e5b-2f6b-4ad4-aa14-8736ec9f4f3a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.854031 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" event={"ID":"e7e9748a-1d74-4b39-9bf0-5c94dda32218","Type":"ContainerStarted","Data":"20e225726cb8690fef73f681e053f73bf9cdd1d6c0813a3e790395c81fac2e47"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.859769 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" event={"ID":"45e13443-0823-411e-bac1-10eadf9b9918","Type":"ContainerStarted","Data":"bd36336f5afd74347e651a0e445b59ece330549c5b058d69b50ad8d8dc1121a7"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.859818 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" event={"ID":"45e13443-0823-411e-bac1-10eadf9b9918","Type":"ContainerStarted","Data":"bcf11de82923f6e2060be87aa31ad1b6bdab502c60769fddac8b3f4945ceba35"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.860646 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.874933 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-l7dgh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.875001 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" podUID="45e13443-0823-411e-bac1-10eadf9b9918" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.876184 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.876564 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.376544973 +0000 UTC m=+155.767521531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.876939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" event={"ID":"9864b4d8-c2ab-44ca-8d2b-cca19b047185","Type":"ContainerStarted","Data":"0aac805343713aa377bd389c972b770e51e3ec8ac53b168906e2be6306b30456"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.876976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" event={"ID":"9864b4d8-c2ab-44ca-8d2b-cca19b047185","Type":"ContainerStarted","Data":"d2c789428d07117c7171907086d17886b0a0478c7c0508c96160baf726a314be"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.892234 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" event={"ID":"8f413f0a-b891-46aa-975f-388300b1931d","Type":"ContainerStarted","Data":"9a20fd1b1baf5cad26e77e5c0f7d7a393dfc7c7f7047b9b38f2c4f210456bec6"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.914674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" event={"ID":"2c80d5ba-231a-4567-874c-116b797c940e","Type":"ContainerStarted","Data":"336fb7b3eb223085bb521279d8e33cbe3cb930784cae72347445883b4d950d99"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.914727 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" event={"ID":"2c80d5ba-231a-4567-874c-116b797c940e","Type":"ContainerStarted","Data":"8e74ce08596219bad8023b6db7d127366d0a799db9b7cf8a678a022d4a8a7370"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.944875 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" event={"ID":"f8904696-72b4-4d00-83f1-552b9909afcf","Type":"ContainerStarted","Data":"6e8a16876cc38a28e4935fbb693cecf38c359c3f01c45f334813ae22e507d50e"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.944923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" event={"ID":"f8904696-72b4-4d00-83f1-552b9909afcf","Type":"ContainerStarted","Data":"13d168d5409bcd2dcbcd0bbb3e8294dcdc0b8c2f5a074e6d871b9c8f61311ea4"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.974930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" event={"ID":"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6","Type":"ContainerStarted","Data":"cee4822df0a01652c6c9529b375ff4cda979f5cccb5368038d13a081ea88547a"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.974994 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" event={"ID":"ddd951b5-5bbf-4f65-b09a-4482bb58fcb6","Type":"ContainerStarted","Data":"b152e19a193954ede1eae84125967c5ba74038822589a559082af93477dc9aa2"} Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.975623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.981697 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:10:59 crc kubenswrapper[4809]: E1205 11:10:59.985982 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.485964211 +0000 UTC m=+155.876940769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:10:59 crc kubenswrapper[4809]: I1205 11:10:59.991025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" event={"ID":"b7406036-f795-40ba-b663-6b17adeabc8a","Type":"ContainerStarted","Data":"7b2f8b0772eded774f8d213c4c94c1312895ab67737282b3ea5ed417de6a6878"} Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.035194 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" event={"ID":"f0cdbeff-6bd6-4454-b0c4-1d556e43cf89","Type":"ContainerStarted","Data":"0e0cf95ffb22490fe9add07340cc7112a8d89157b72bebdff8c14566f5181ba6"} Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.035726 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.047024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" event={"ID":"d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2","Type":"ContainerStarted","Data":"f2ea36805a58c735f51276459dc1da8fddd4527872247be973b401063895a4d8"} Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.061336 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bvjsr" podStartSLOduration=132.061318288 podStartE2EDuration="2m12.061318288s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.044677238 +0000 UTC m=+155.435653806" watchObservedRunningTime="2025-12-05 11:11:00.061318288 +0000 UTC m=+155.452294846" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.082536 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.083537 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.583520884 +0000 UTC m=+155.974497442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.086773 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" event={"ID":"520f140b-3326-4ec2-8a13-a1142581c831","Type":"ContainerStarted","Data":"8b82b7eff5aff700eda56dff90d95b6cdc8b648572c0a8f0c46ec022b9c23c3d"} Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.090994 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wfbxx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.091057 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.091142 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.091158 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.181820 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:00 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:00 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:00 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.182088 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.186495 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.194345 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.694333972 +0000 UTC m=+156.085310530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.290140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.314997 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.814976746 +0000 UTC m=+156.205953304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.338235 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" podStartSLOduration=132.338213382 podStartE2EDuration="2m12.338213382s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.336186935 +0000 UTC m=+155.727163513" watchObservedRunningTime="2025-12-05 11:11:00.338213382 +0000 UTC m=+155.729189940" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.415670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.415922 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:00.915910564 +0000 UTC m=+156.306887122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.437542 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.448855 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" podStartSLOduration=132.448828094 podStartE2EDuration="2m12.448828094s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.448786082 +0000 UTC m=+155.839762640" watchObservedRunningTime="2025-12-05 11:11:00.448828094 +0000 UTC m=+155.839804652" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.545573 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.546044 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.046027276 +0000 UTC m=+156.437003834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.562605 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" podStartSLOduration=132.562585904 podStartE2EDuration="2m12.562585904s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.561159284 +0000 UTC m=+155.952135852" watchObservedRunningTime="2025-12-05 11:11:00.562585904 +0000 UTC m=+155.953562472" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.647720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.648090 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.148076496 +0000 UTC m=+156.539053054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.748023 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-wml6l" podStartSLOduration=9.748006877 podStartE2EDuration="9.748006877s" podCreationTimestamp="2025-12-05 11:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.699823277 +0000 UTC m=+156.090799835" watchObservedRunningTime="2025-12-05 11:11:00.748006877 +0000 UTC m=+156.138983435" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.748586 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rxf5n" podStartSLOduration=132.748580693 podStartE2EDuration="2m12.748580693s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.747515893 +0000 UTC m=+156.138492471" watchObservedRunningTime="2025-12-05 11:11:00.748580693 +0000 UTC m=+156.139557251" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.748886 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.749210 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.24919253 +0000 UTC m=+156.640169088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.856348 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.856771 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.356754936 +0000 UTC m=+156.747731494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.859950 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-shlh8" podStartSLOduration=132.859929375 podStartE2EDuration="2m12.859929375s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.857843476 +0000 UTC m=+156.248820034" watchObservedRunningTime="2025-12-05 11:11:00.859929375 +0000 UTC m=+156.250905933" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.862621 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fhht6" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.892272 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" podStartSLOduration=132.892256558 podStartE2EDuration="2m12.892256558s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.890400475 +0000 UTC m=+156.281377033" watchObservedRunningTime="2025-12-05 11:11:00.892256558 +0000 UTC m=+156.283233116" Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.957705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:00 crc kubenswrapper[4809]: E1205 11:11:00.957948 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.457931441 +0000 UTC m=+156.848907999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:00 crc kubenswrapper[4809]: I1205 11:11:00.987837 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9b4bq" podStartSLOduration=132.987818354 podStartE2EDuration="2m12.987818354s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:00.987079703 +0000 UTC m=+156.378056281" watchObservedRunningTime="2025-12-05 11:11:00.987818354 +0000 UTC m=+156.378794912" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.066316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.067859 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.567844533 +0000 UTC m=+156.958821091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.128104 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" event={"ID":"f8904696-72b4-4d00-83f1-552b9909afcf","Type":"ContainerStarted","Data":"96a098af52dfc8a16bbb7924202afdb0ed6b57f932abde209d3ddc4f74d45465"} Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.130668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" event={"ID":"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7","Type":"ContainerStarted","Data":"b04d65e742487cff5d708f5aafe56f13b9e65520d00de17637484c2c3976712f"} Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.140843 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wfbxx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.140851 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-l7dgh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.140921 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.140958 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" podUID="45e13443-0823-411e-bac1-10eadf9b9918" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.141132 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.141185 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.161192 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5rqh4" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.164609 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:01 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:01 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:01 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.164669 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.167294 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.167606 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.667577827 +0000 UTC m=+157.058554385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.201199 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-bcsqv" podStartSLOduration=133.201174356 podStartE2EDuration="2m13.201174356s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:01.128987298 +0000 UTC m=+156.519963866" watchObservedRunningTime="2025-12-05 11:11:01.201174356 +0000 UTC m=+156.592150914" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.204572 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.205327 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.227337 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmv58" podStartSLOduration=133.227320603 podStartE2EDuration="2m13.227320603s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:01.207625458 +0000 UTC m=+156.598602016" watchObservedRunningTime="2025-12-05 11:11:01.227320603 +0000 UTC m=+156.618297161" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.228340 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.228526 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.355699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.356031 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.856011925 +0000 UTC m=+157.246988493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.361850 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.384848 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-8d57d" podStartSLOduration=10.384821138 podStartE2EDuration="10.384821138s" podCreationTimestamp="2025-12-05 11:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:01.375734392 +0000 UTC m=+156.766710950" watchObservedRunningTime="2025-12-05 11:11:01.384821138 +0000 UTC m=+156.775797696" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.462049 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.462225 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.462288 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.462420 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:01.962403068 +0000 UTC m=+157.353379626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.469264 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" podStartSLOduration=133.469248991 podStartE2EDuration="2m13.469248991s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:01.467926094 +0000 UTC m=+156.858902672" watchObservedRunningTime="2025-12-05 11:11:01.469248991 +0000 UTC m=+156.860225549" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.469764 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bstzg" podStartSLOduration=133.469760505 podStartE2EDuration="2m13.469760505s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:01.422361518 +0000 UTC m=+156.813338076" watchObservedRunningTime="2025-12-05 11:11:01.469760505 +0000 UTC m=+156.860737063" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.563188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.563274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.563321 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.563706 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.564000 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.063987505 +0000 UTC m=+157.454964063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.595011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.664421 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.665106 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.165086107 +0000 UTC m=+157.556062665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.719941 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pbsgg" podStartSLOduration=133.719920404 podStartE2EDuration="2m13.719920404s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:01.689592288 +0000 UTC m=+157.080568846" watchObservedRunningTime="2025-12-05 11:11:01.719920404 +0000 UTC m=+157.110896962" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.766763 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.767143 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.267131937 +0000 UTC m=+157.658108495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.830585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.867597 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.867750 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.367732756 +0000 UTC m=+157.758709314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.867836 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.868222 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.368212739 +0000 UTC m=+157.759189297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:01 crc kubenswrapper[4809]: I1205 11:11:01.969592 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:01 crc kubenswrapper[4809]: E1205 11:11:01.969879 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.469862038 +0000 UTC m=+157.860838596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.075399 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.075737 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.575722785 +0000 UTC m=+157.966699343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.171948 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:02 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:02 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:02 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.172011 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.176030 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.176448 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.676428687 +0000 UTC m=+158.067405245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.259117 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" event={"ID":"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7","Type":"ContainerStarted","Data":"8e17b01c03abac496fa325bef10be6d7023489cf60bffb4fe2c68039f0c089e3"} Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.260787 4809 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-l7dgh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.260842 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" podUID="45e13443-0823-411e-bac1-10eadf9b9918" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.285835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.286584 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.786573676 +0000 UTC m=+158.177550234 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.295678 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rwmwd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.295748 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" podUID="f0cdbeff-6bd6-4454-b0c4-1d556e43cf89" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.387449 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.387692 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.887662628 +0000 UTC m=+158.278639186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.387753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.388017 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.888004928 +0000 UTC m=+158.278981486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.488638 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.488799 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.988771602 +0000 UTC m=+158.379748160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.488933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.489217 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:02.989205794 +0000 UTC m=+158.380182352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.591379 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.591933 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.091918533 +0000 UTC m=+158.482895091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.679683 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.679715 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.695125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.696378 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.19636279 +0000 UTC m=+158.587339348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.708017 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.708052 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.740823 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.767431 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pf4fm" podStartSLOduration=134.767406285 podStartE2EDuration="2m14.767406285s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:01.72226936 +0000 UTC m=+157.113245918" watchObservedRunningTime="2025-12-05 11:11:02.767406285 +0000 UTC m=+158.158382843" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.769838 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.770490 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.780989 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.785084 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.786414 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.800829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.801143 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/114480e0-db52-4655-a6e9-7477e5aa7e0f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.801206 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/114480e0-db52-4655-a6e9-7477e5aa7e0f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.801925 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.301907009 +0000 UTC m=+158.692883567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.817461 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rwmwd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.817516 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" podUID="f0cdbeff-6bd6-4454-b0c4-1d556e43cf89" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.818094 4809 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-rwmwd container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.818125 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" podUID="f0cdbeff-6bd6-4454-b0c4-1d556e43cf89" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.853339 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 11:11:02 crc kubenswrapper[4809]: W1205 11:11:02.875686 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7958a3c3_d6f2_4d04_a773_8bf0ceb0e187.slice/crio-825216fcb50cf40784054fdf5fab7c5ad18d544a50cb6bfb1a00be05d1bfdd26 WatchSource:0}: Error finding container 825216fcb50cf40784054fdf5fab7c5ad18d544a50cb6bfb1a00be05d1bfdd26: Status 404 returned error can't find the container with id 825216fcb50cf40784054fdf5fab7c5ad18d544a50cb6bfb1a00be05d1bfdd26 Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.903093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/114480e0-db52-4655-a6e9-7477e5aa7e0f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.903156 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/114480e0-db52-4655-a6e9-7477e5aa7e0f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.903216 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:02 crc kubenswrapper[4809]: E1205 11:11:02.903563 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.403548437 +0000 UTC m=+158.794524995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.905245 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/114480e0-db52-4655-a6e9-7477e5aa7e0f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:02 crc kubenswrapper[4809]: I1205 11:11:02.932284 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/114480e0-db52-4655-a6e9-7477e5aa7e0f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.004055 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.004340 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.504323461 +0000 UTC m=+158.895300009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.105578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.106207 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.606186116 +0000 UTC m=+158.997162704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.117061 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.166224 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:03 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:03 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:03 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.166289 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.172352 4809 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kpt8q container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]log ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]etcd ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/generic-apiserver-start-informers ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/max-in-flight-filter ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 05 11:11:03 crc kubenswrapper[4809]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 05 11:11:03 crc kubenswrapper[4809]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/project.openshift.io-projectcache ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/openshift.io-startinformers ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 05 11:11:03 crc kubenswrapper[4809]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 05 11:11:03 crc kubenswrapper[4809]: livez check failed Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.172414 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" podUID="d34ae5bd-80d8-48d3-a0ea-55407b8ee8f2" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.206482 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.206771 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.706730643 +0000 UTC m=+159.097707201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.207194 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.207596 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.707581057 +0000 UTC m=+159.098557625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.271525 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187","Type":"ContainerStarted","Data":"825216fcb50cf40784054fdf5fab7c5ad18d544a50cb6bfb1a00be05d1bfdd26"} Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.316146 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-w7ffm" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.316620 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.316983 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.816967854 +0000 UTC m=+159.207944412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.426077 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.427883 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:03.927868894 +0000 UTC m=+159.318845452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.529116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.529685 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.029668927 +0000 UTC m=+159.420645475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.535200 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 11:11:03 crc kubenswrapper[4809]: W1205 11:11:03.575065 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod114480e0_db52_4655_a6e9_7477e5aa7e0f.slice/crio-af154aedcaf4016df3c7cd980ab1e8bfd20f8a3f9f89378d5db823351f791b14 WatchSource:0}: Error finding container af154aedcaf4016df3c7cd980ab1e8bfd20f8a3f9f89378d5db823351f791b14: Status 404 returned error can't find the container with id af154aedcaf4016df3c7cd980ab1e8bfd20f8a3f9f89378d5db823351f791b14 Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.614694 4809 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.635304 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.635712 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.135698359 +0000 UTC m=+159.526674927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.736592 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.737021 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.236999078 +0000 UTC m=+159.627975646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.841758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.842094 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.342081774 +0000 UTC m=+159.733058342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.865114 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.865146 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.865161 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.865190 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.871458 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.871588 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.873241 4809 patch_prober.go:28] interesting pod/console-f9d7485db-zhtzr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.873279 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zhtzr" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.942729 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.942910 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.442882638 +0000 UTC m=+159.833859186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:03 crc kubenswrapper[4809]: I1205 11:11:03.943014 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:03 crc kubenswrapper[4809]: E1205 11:11:03.944239 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.444220006 +0000 UTC m=+159.835196574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.044072 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:04 crc kubenswrapper[4809]: E1205 11:11:04.044497 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.544478405 +0000 UTC m=+159.935454963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.145190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:04 crc kubenswrapper[4809]: E1205 11:11:04.145559 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 11:11:04.645548457 +0000 UTC m=+160.036525015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mx7gs" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.160716 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.163184 4809 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-05T11:11:03.614729197Z","Handler":null,"Name":""} Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.164433 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:04 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:04 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:04 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.164485 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.182596 4809 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.182649 4809 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.245869 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.262944 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.280426 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187","Type":"ContainerStarted","Data":"a92313896ee370cc1cf743e8cd13e7a8d321cd3b5aa42d85236f40c924dd385c"} Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.282357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"114480e0-db52-4655-a6e9-7477e5aa7e0f","Type":"ContainerStarted","Data":"af154aedcaf4016df3c7cd980ab1e8bfd20f8a3f9f89378d5db823351f791b14"} Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.285368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" event={"ID":"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7","Type":"ContainerStarted","Data":"6b0b8a9ab01a2d954b5042055000c79174aa594dc98fe1e5d8c1d3fc93801b3a"} Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.347588 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.370644 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.370684 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.404846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mx7gs\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.417714 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.432182 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.432163156 podStartE2EDuration="3.432163156s" podCreationTimestamp="2025-12-05 11:11:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:04.29872258 +0000 UTC m=+159.689699138" watchObservedRunningTime="2025-12-05 11:11:04.432163156 +0000 UTC m=+159.823139714" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.474912 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l7dgh" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.554105 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:04 crc kubenswrapper[4809]: I1205 11:11:04.899670 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.025157 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9t5j4"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.026025 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.028789 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9t5j4"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.029057 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.077300 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx7gs"] Dec 05 11:11:05 crc kubenswrapper[4809]: W1205 11:11:05.083415 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1de639b4_f614_41b9_812b_a7bfea3ccc90.slice/crio-8587c79bbc94e25ef6c22c3a9eab8bc8faf7580d3dbc2bfc1a9afb80f966a492 WatchSource:0}: Error finding container 8587c79bbc94e25ef6c22c3a9eab8bc8faf7580d3dbc2bfc1a9afb80f966a492: Status 404 returned error can't find the container with id 8587c79bbc94e25ef6c22c3a9eab8bc8faf7580d3dbc2bfc1a9afb80f966a492 Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.158387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxjwx\" (UniqueName: \"kubernetes.io/projected/75df67eb-0996-4396-94bb-8ddbce4136ce-kube-api-access-nxjwx\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.158430 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-catalog-content\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.158619 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-utilities\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.163173 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:05 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:05 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:05 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.163246 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.217539 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xj244"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.218848 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.221055 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.229995 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xj244"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.259806 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-utilities\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.260031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxjwx\" (UniqueName: \"kubernetes.io/projected/75df67eb-0996-4396-94bb-8ddbce4136ce-kube-api-access-nxjwx\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.260089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-catalog-content\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.260316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-utilities\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.260509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-catalog-content\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.286918 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxjwx\" (UniqueName: \"kubernetes.io/projected/75df67eb-0996-4396-94bb-8ddbce4136ce-kube-api-access-nxjwx\") pod \"certified-operators-9t5j4\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.291736 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" event={"ID":"392dbe41-06c4-45cd-9e24-69ecc8f1b8e7","Type":"ContainerStarted","Data":"d5e2d032aa4b96ffed77d8db46cdaef67a56e3726ff94c0a179f8c5c78b2c5b9"} Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.293958 4809 generic.go:334] "Generic (PLEG): container finished" podID="7958a3c3-d6f2-4d04-a773-8bf0ceb0e187" containerID="a92313896ee370cc1cf743e8cd13e7a8d321cd3b5aa42d85236f40c924dd385c" exitCode=0 Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.294024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187","Type":"ContainerDied","Data":"a92313896ee370cc1cf743e8cd13e7a8d321cd3b5aa42d85236f40c924dd385c"} Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.295165 4809 generic.go:334] "Generic (PLEG): container finished" podID="114480e0-db52-4655-a6e9-7477e5aa7e0f" containerID="f764c3cab37b69821fea626b1ef71bfc4e444e41064637faed589ba0f6b394d9" exitCode=0 Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.295288 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"114480e0-db52-4655-a6e9-7477e5aa7e0f","Type":"ContainerDied","Data":"f764c3cab37b69821fea626b1ef71bfc4e444e41064637faed589ba0f6b394d9"} Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.297026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" event={"ID":"1de639b4-f614-41b9-812b-a7bfea3ccc90","Type":"ContainerStarted","Data":"8587c79bbc94e25ef6c22c3a9eab8bc8faf7580d3dbc2bfc1a9afb80f966a492"} Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.300213 4809 generic.go:334] "Generic (PLEG): container finished" podID="e7e9748a-1d74-4b39-9bf0-5c94dda32218" containerID="20e225726cb8690fef73f681e053f73bf9cdd1d6c0813a3e790395c81fac2e47" exitCode=0 Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.300258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" event={"ID":"e7e9748a-1d74-4b39-9bf0-5c94dda32218","Type":"ContainerDied","Data":"20e225726cb8690fef73f681e053f73bf9cdd1d6c0813a3e790395c81fac2e47"} Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.322103 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2rpwq" podStartSLOduration=14.32208474 podStartE2EDuration="14.32208474s" podCreationTimestamp="2025-12-05 11:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:05.320020212 +0000 UTC m=+160.710996800" watchObservedRunningTime="2025-12-05 11:11:05.32208474 +0000 UTC m=+160.713061308" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.347541 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.363141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m9b7\" (UniqueName: \"kubernetes.io/projected/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-kube-api-access-6m9b7\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.363188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-utilities\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.363288 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-catalog-content\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.425272 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lfv24"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.426731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.434970 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lfv24"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.464665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m9b7\" (UniqueName: \"kubernetes.io/projected/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-kube-api-access-6m9b7\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.464720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-utilities\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.464786 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-catalog-content\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.464807 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-utilities\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.464837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqkrs\" (UniqueName: \"kubernetes.io/projected/50cea908-3bdc-4a86-b577-47fa705d1957-kube-api-access-wqkrs\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.464860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-catalog-content\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.465298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-catalog-content\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.465815 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-utilities\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.483389 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m9b7\" (UniqueName: \"kubernetes.io/projected/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-kube-api-access-6m9b7\") pod \"community-operators-xj244\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.565243 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-catalog-content\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.565284 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-utilities\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.565314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqkrs\" (UniqueName: \"kubernetes.io/projected/50cea908-3bdc-4a86-b577-47fa705d1957-kube-api-access-wqkrs\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.566532 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-catalog-content\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.567521 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-utilities\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.567795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.586308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqkrs\" (UniqueName: \"kubernetes.io/projected/50cea908-3bdc-4a86-b577-47fa705d1957-kube-api-access-wqkrs\") pod \"certified-operators-lfv24\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.650870 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-chdkw"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.655494 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.655933 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-chdkw"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.750192 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.771291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkqqs\" (UniqueName: \"kubernetes.io/projected/4bfa41b4-f65e-49a2-9b3c-783194dc934d-kube-api-access-fkqqs\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.771342 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-utilities\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.771381 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-catalog-content\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.808303 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.812349 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9t5j4"] Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.839436 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwmwd" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.876665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-utilities\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.876824 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-catalog-content\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.876973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkqqs\" (UniqueName: \"kubernetes.io/projected/4bfa41b4-f65e-49a2-9b3c-783194dc934d-kube-api-access-fkqqs\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.877767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-utilities\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.879037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-catalog-content\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.913037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkqqs\" (UniqueName: \"kubernetes.io/projected/4bfa41b4-f65e-49a2-9b3c-783194dc934d-kube-api-access-fkqqs\") pod \"community-operators-chdkw\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:05 crc kubenswrapper[4809]: I1205 11:11:05.994943 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.102461 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xj244"] Dec 05 11:11:06 crc kubenswrapper[4809]: W1205 11:11:06.113388 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aaa8ff8_bbd1_4768_9325_285eb8acb01a.slice/crio-7cca5619e65eafd6f3e9e1fdeefb4680bb5309e3638a10b4e42a29ae6cf3e58b WatchSource:0}: Error finding container 7cca5619e65eafd6f3e9e1fdeefb4680bb5309e3638a10b4e42a29ae6cf3e58b: Status 404 returned error can't find the container with id 7cca5619e65eafd6f3e9e1fdeefb4680bb5309e3638a10b4e42a29ae6cf3e58b Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.181370 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:06 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:06 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:06 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.181409 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.188310 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lfv24"] Dec 05 11:11:06 crc kubenswrapper[4809]: W1205 11:11:06.211489 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50cea908_3bdc_4a86_b577_47fa705d1957.slice/crio-afea6276daff65f99c186ef5de375ca6b39325febd4b18885e34d3094f90a3f9 WatchSource:0}: Error finding container afea6276daff65f99c186ef5de375ca6b39325febd4b18885e34d3094f90a3f9: Status 404 returned error can't find the container with id afea6276daff65f99c186ef5de375ca6b39325febd4b18885e34d3094f90a3f9 Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.313008 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xj244" event={"ID":"5aaa8ff8-bbd1-4768-9325-285eb8acb01a","Type":"ContainerStarted","Data":"7cca5619e65eafd6f3e9e1fdeefb4680bb5309e3638a10b4e42a29ae6cf3e58b"} Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.317677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" event={"ID":"1de639b4-f614-41b9-812b-a7bfea3ccc90","Type":"ContainerStarted","Data":"824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf"} Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.317838 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.320253 4809 generic.go:334] "Generic (PLEG): container finished" podID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerID="4fb36aaccd438d4f57c7e7a19206bfca03603bf8381db1d876e9131355a0d985" exitCode=0 Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.320329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9t5j4" event={"ID":"75df67eb-0996-4396-94bb-8ddbce4136ce","Type":"ContainerDied","Data":"4fb36aaccd438d4f57c7e7a19206bfca03603bf8381db1d876e9131355a0d985"} Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.320358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9t5j4" event={"ID":"75df67eb-0996-4396-94bb-8ddbce4136ce","Type":"ContainerStarted","Data":"0fab17bc7202aac0e33e18c4e3718817ee6309197f48955b3c35cd1e4238ee62"} Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.321885 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.322884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfv24" event={"ID":"50cea908-3bdc-4a86-b577-47fa705d1957","Type":"ContainerStarted","Data":"afea6276daff65f99c186ef5de375ca6b39325febd4b18885e34d3094f90a3f9"} Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.357157 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" podStartSLOduration=138.35713418 podStartE2EDuration="2m18.35713418s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:11:06.353323102 +0000 UTC m=+161.744299660" watchObservedRunningTime="2025-12-05 11:11:06.35713418 +0000 UTC m=+161.748110738" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.372714 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-chdkw"] Dec 05 11:11:06 crc kubenswrapper[4809]: E1205 11:11:06.455515 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aaa8ff8_bbd1_4768_9325_285eb8acb01a.slice/crio-conmon-c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aaa8ff8_bbd1_4768_9325_285eb8acb01a.slice/crio-c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf.scope\": RecentStats: unable to find data in memory cache]" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.653262 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.690331 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.753877 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.789115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7e9748a-1d74-4b39-9bf0-5c94dda32218-config-volume\") pod \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.789264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69dxb\" (UniqueName: \"kubernetes.io/projected/e7e9748a-1d74-4b39-9bf0-5c94dda32218-kube-api-access-69dxb\") pod \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.789300 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/114480e0-db52-4655-a6e9-7477e5aa7e0f-kubelet-dir\") pod \"114480e0-db52-4655-a6e9-7477e5aa7e0f\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.789337 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/114480e0-db52-4655-a6e9-7477e5aa7e0f-kube-api-access\") pod \"114480e0-db52-4655-a6e9-7477e5aa7e0f\" (UID: \"114480e0-db52-4655-a6e9-7477e5aa7e0f\") " Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.789367 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7e9748a-1d74-4b39-9bf0-5c94dda32218-secret-volume\") pod \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\" (UID: \"e7e9748a-1d74-4b39-9bf0-5c94dda32218\") " Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.789739 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/114480e0-db52-4655-a6e9-7477e5aa7e0f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "114480e0-db52-4655-a6e9-7477e5aa7e0f" (UID: "114480e0-db52-4655-a6e9-7477e5aa7e0f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.790177 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e9748a-1d74-4b39-9bf0-5c94dda32218-config-volume" (OuterVolumeSpecName: "config-volume") pod "e7e9748a-1d74-4b39-9bf0-5c94dda32218" (UID: "e7e9748a-1d74-4b39-9bf0-5c94dda32218"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.794366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/114480e0-db52-4655-a6e9-7477e5aa7e0f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "114480e0-db52-4655-a6e9-7477e5aa7e0f" (UID: "114480e0-db52-4655-a6e9-7477e5aa7e0f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.794696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e9748a-1d74-4b39-9bf0-5c94dda32218-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e7e9748a-1d74-4b39-9bf0-5c94dda32218" (UID: "e7e9748a-1d74-4b39-9bf0-5c94dda32218"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.794844 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e9748a-1d74-4b39-9bf0-5c94dda32218-kube-api-access-69dxb" (OuterVolumeSpecName: "kube-api-access-69dxb") pod "e7e9748a-1d74-4b39-9bf0-5c94dda32218" (UID: "e7e9748a-1d74-4b39-9bf0-5c94dda32218"). InnerVolumeSpecName "kube-api-access-69dxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890617 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kubelet-dir\") pod \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890675 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kube-api-access\") pod \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\" (UID: \"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187\") " Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890742 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7958a3c3-d6f2-4d04-a773-8bf0ceb0e187" (UID: "7958a3c3-d6f2-4d04-a773-8bf0ceb0e187"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890878 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69dxb\" (UniqueName: \"kubernetes.io/projected/e7e9748a-1d74-4b39-9bf0-5c94dda32218-kube-api-access-69dxb\") on node \"crc\" DevicePath \"\"" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890892 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/114480e0-db52-4655-a6e9-7477e5aa7e0f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890900 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/114480e0-db52-4655-a6e9-7477e5aa7e0f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890909 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7e9748a-1d74-4b39-9bf0-5c94dda32218-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890917 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7e9748a-1d74-4b39-9bf0-5c94dda32218-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.890925 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.893956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7958a3c3-d6f2-4d04-a773-8bf0ceb0e187" (UID: "7958a3c3-d6f2-4d04-a773-8bf0ceb0e187"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:11:06 crc kubenswrapper[4809]: I1205 11:11:06.992122 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7958a3c3-d6f2-4d04-a773-8bf0ceb0e187-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.022230 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qcp4f"] Dec 05 11:11:07 crc kubenswrapper[4809]: E1205 11:11:07.022462 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e9748a-1d74-4b39-9bf0-5c94dda32218" containerName="collect-profiles" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.022477 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e9748a-1d74-4b39-9bf0-5c94dda32218" containerName="collect-profiles" Dec 05 11:11:07 crc kubenswrapper[4809]: E1205 11:11:07.022496 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="114480e0-db52-4655-a6e9-7477e5aa7e0f" containerName="pruner" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.022505 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="114480e0-db52-4655-a6e9-7477e5aa7e0f" containerName="pruner" Dec 05 11:11:07 crc kubenswrapper[4809]: E1205 11:11:07.022517 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7958a3c3-d6f2-4d04-a773-8bf0ceb0e187" containerName="pruner" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.022526 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7958a3c3-d6f2-4d04-a773-8bf0ceb0e187" containerName="pruner" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.022701 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e9748a-1d74-4b39-9bf0-5c94dda32218" containerName="collect-profiles" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.022730 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7958a3c3-d6f2-4d04-a773-8bf0ceb0e187" containerName="pruner" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.022750 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="114480e0-db52-4655-a6e9-7477e5aa7e0f" containerName="pruner" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.023915 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.026115 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.034878 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcp4f"] Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.162720 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:07 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:07 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:07 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.162776 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.193965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-catalog-content\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.194010 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-utilities\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.194034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hq4g\" (UniqueName: \"kubernetes.io/projected/29be0f64-8b6e-453e-a8c0-dbe464da2705-kube-api-access-9hq4g\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.300376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hq4g\" (UniqueName: \"kubernetes.io/projected/29be0f64-8b6e-453e-a8c0-dbe464da2705-kube-api-access-9hq4g\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.300491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-catalog-content\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.300519 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-utilities\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.300957 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-utilities\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.301160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-catalog-content\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.318455 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hq4g\" (UniqueName: \"kubernetes.io/projected/29be0f64-8b6e-453e-a8c0-dbe464da2705-kube-api-access-9hq4g\") pod \"redhat-marketplace-qcp4f\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.328571 4809 generic.go:334] "Generic (PLEG): container finished" podID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerID="c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf" exitCode=0 Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.328661 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xj244" event={"ID":"5aaa8ff8-bbd1-4768-9325-285eb8acb01a","Type":"ContainerDied","Data":"c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf"} Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.330799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chdkw" event={"ID":"4bfa41b4-f65e-49a2-9b3c-783194dc934d","Type":"ContainerDied","Data":"2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993"} Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.330697 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerID="2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993" exitCode=0 Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.330948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chdkw" event={"ID":"4bfa41b4-f65e-49a2-9b3c-783194dc934d","Type":"ContainerStarted","Data":"e39306b9f3d349ec761e00f33607cbedfbc07b459df76f2f855a6f33f9bbee73"} Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.332811 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.332850 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8" event={"ID":"e7e9748a-1d74-4b39-9bf0-5c94dda32218","Type":"ContainerDied","Data":"034525a693eef5658a15b496f3a8691b5c169f3a632261af5bfd075fccfc87b3"} Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.332883 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="034525a693eef5658a15b496f3a8691b5c169f3a632261af5bfd075fccfc87b3" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.337795 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.337798 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7958a3c3-d6f2-4d04-a773-8bf0ceb0e187","Type":"ContainerDied","Data":"825216fcb50cf40784054fdf5fab7c5ad18d544a50cb6bfb1a00be05d1bfdd26"} Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.337830 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="825216fcb50cf40784054fdf5fab7c5ad18d544a50cb6bfb1a00be05d1bfdd26" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.340097 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.340177 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"114480e0-db52-4655-a6e9-7477e5aa7e0f","Type":"ContainerDied","Data":"af154aedcaf4016df3c7cd980ab1e8bfd20f8a3f9f89378d5db823351f791b14"} Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.340231 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af154aedcaf4016df3c7cd980ab1e8bfd20f8a3f9f89378d5db823351f791b14" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.343132 4809 generic.go:334] "Generic (PLEG): container finished" podID="50cea908-3bdc-4a86-b577-47fa705d1957" containerID="494545a5aca2d31affd87654fa17184d9898bfec1b700ccff272628697713457" exitCode=0 Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.343213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfv24" event={"ID":"50cea908-3bdc-4a86-b577-47fa705d1957","Type":"ContainerDied","Data":"494545a5aca2d31affd87654fa17184d9898bfec1b700ccff272628697713457"} Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.344563 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.540412 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dmzbq"] Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.544105 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.554386 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmzbq"] Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.604319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xbbf\" (UniqueName: \"kubernetes.io/projected/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-kube-api-access-4xbbf\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.604421 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-catalog-content\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.604470 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-utilities\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.684485 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.689497 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kpt8q" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.707796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xbbf\" (UniqueName: \"kubernetes.io/projected/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-kube-api-access-4xbbf\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.707845 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-catalog-content\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.707877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-utilities\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.709529 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-catalog-content\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.709827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-utilities\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.786104 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xbbf\" (UniqueName: \"kubernetes.io/projected/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-kube-api-access-4xbbf\") pod \"redhat-marketplace-dmzbq\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:07 crc kubenswrapper[4809]: I1205 11:11:07.864455 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.015380 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcp4f"] Dec 05 11:11:08 crc kubenswrapper[4809]: W1205 11:11:08.034360 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29be0f64_8b6e_453e_a8c0_dbe464da2705.slice/crio-1e2301b3a09c3dd63861051666265c2ced971e2fd192ce1c5c235aceb8db777e WatchSource:0}: Error finding container 1e2301b3a09c3dd63861051666265c2ced971e2fd192ce1c5c235aceb8db777e: Status 404 returned error can't find the container with id 1e2301b3a09c3dd63861051666265c2ced971e2fd192ce1c5c235aceb8db777e Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.164549 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:08 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:08 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:08 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.164618 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.324789 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmzbq"] Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.351852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcp4f" event={"ID":"29be0f64-8b6e-453e-a8c0-dbe464da2705","Type":"ContainerStarted","Data":"1e2301b3a09c3dd63861051666265c2ced971e2fd192ce1c5c235aceb8db777e"} Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.353760 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmzbq" event={"ID":"73486c5f-7eeb-4a30-ad5c-2eaf64837d28","Type":"ContainerStarted","Data":"8ba555b4d9ceff40801fd9fdad25448b1e55c2b567ea41cd22ec0b468e547785"} Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.432470 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jrg7v"] Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.433686 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.436684 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.449852 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jrg7v"] Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.467687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-catalog-content\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.467845 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-utilities\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.467886 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrxvr\" (UniqueName: \"kubernetes.io/projected/e93c2521-42a0-4944-97c7-448e498294a4-kube-api-access-wrxvr\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.568771 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-catalog-content\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.568860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-utilities\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.568887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrxvr\" (UniqueName: \"kubernetes.io/projected/e93c2521-42a0-4944-97c7-448e498294a4-kube-api-access-wrxvr\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.569799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-utilities\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.570036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-catalog-content\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.592733 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrxvr\" (UniqueName: \"kubernetes.io/projected/e93c2521-42a0-4944-97c7-448e498294a4-kube-api-access-wrxvr\") pod \"redhat-operators-jrg7v\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.773705 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.819177 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-wml6l" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.820175 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hlxqf"] Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.821777 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.836177 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hlxqf"] Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.871832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-catalog-content\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.871880 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/b6f405ac-b705-477f-b150-109df1092b2b-kube-api-access-gdfjz\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.871915 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-utilities\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.973655 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/b6f405ac-b705-477f-b150-109df1092b2b-kube-api-access-gdfjz\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.973713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-catalog-content\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.973749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-utilities\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.974354 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-utilities\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.975130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-catalog-content\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:08 crc kubenswrapper[4809]: I1205 11:11:08.998746 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/b6f405ac-b705-477f-b150-109df1092b2b-kube-api-access-gdfjz\") pod \"redhat-operators-hlxqf\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.141869 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jrg7v"] Dec 05 11:11:09 crc kubenswrapper[4809]: W1205 11:11:09.158875 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode93c2521_42a0_4944_97c7_448e498294a4.slice/crio-6af6b0fd47016acf44a939ea845477ed3796925630bbba3127cf7b17c26df7ba WatchSource:0}: Error finding container 6af6b0fd47016acf44a939ea845477ed3796925630bbba3127cf7b17c26df7ba: Status 404 returned error can't find the container with id 6af6b0fd47016acf44a939ea845477ed3796925630bbba3127cf7b17c26df7ba Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.162149 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:09 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:09 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:09 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.162197 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.166998 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.392207 4809 generic.go:334] "Generic (PLEG): container finished" podID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerID="9fae6f55d1ebbf8eacf831d0bef3dc0bc19650b742b754a1cd6851a54d352b6a" exitCode=0 Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.392510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcp4f" event={"ID":"29be0f64-8b6e-453e-a8c0-dbe464da2705","Type":"ContainerDied","Data":"9fae6f55d1ebbf8eacf831d0bef3dc0bc19650b742b754a1cd6851a54d352b6a"} Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.425978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrg7v" event={"ID":"e93c2521-42a0-4944-97c7-448e498294a4","Type":"ContainerStarted","Data":"6af6b0fd47016acf44a939ea845477ed3796925630bbba3127cf7b17c26df7ba"} Dec 05 11:11:09 crc kubenswrapper[4809]: I1205 11:11:09.592009 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hlxqf"] Dec 05 11:11:09 crc kubenswrapper[4809]: W1205 11:11:09.664291 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6f405ac_b705_477f_b150_109df1092b2b.slice/crio-b9545c67a2ad6672c39dbc6bb8267b2f83f9d6c5f09fea92a1c19bce5f159e83 WatchSource:0}: Error finding container b9545c67a2ad6672c39dbc6bb8267b2f83f9d6c5f09fea92a1c19bce5f159e83: Status 404 returned error can't find the container with id b9545c67a2ad6672c39dbc6bb8267b2f83f9d6c5f09fea92a1c19bce5f159e83 Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.162709 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:10 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:10 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:10 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.162758 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.432490 4809 generic.go:334] "Generic (PLEG): container finished" podID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerID="97eeb0e7d0a10e20a743d7895ae6271578c4ef011ec33af5fc054c54eadce737" exitCode=0 Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.432536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmzbq" event={"ID":"73486c5f-7eeb-4a30-ad5c-2eaf64837d28","Type":"ContainerDied","Data":"97eeb0e7d0a10e20a743d7895ae6271578c4ef011ec33af5fc054c54eadce737"} Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.439123 4809 generic.go:334] "Generic (PLEG): container finished" podID="e93c2521-42a0-4944-97c7-448e498294a4" containerID="bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113" exitCode=0 Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.439190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrg7v" event={"ID":"e93c2521-42a0-4944-97c7-448e498294a4","Type":"ContainerDied","Data":"bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113"} Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.441118 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6f405ac-b705-477f-b150-109df1092b2b" containerID="bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca" exitCode=0 Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.441150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxqf" event={"ID":"b6f405ac-b705-477f-b150-109df1092b2b","Type":"ContainerDied","Data":"bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca"} Dec 05 11:11:10 crc kubenswrapper[4809]: I1205 11:11:10.441169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxqf" event={"ID":"b6f405ac-b705-477f-b150-109df1092b2b","Type":"ContainerStarted","Data":"b9545c67a2ad6672c39dbc6bb8267b2f83f9d6c5f09fea92a1c19bce5f159e83"} Dec 05 11:11:11 crc kubenswrapper[4809]: I1205 11:11:11.165877 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:11 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:11 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:11 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:11 crc kubenswrapper[4809]: I1205 11:11:11.165927 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:12 crc kubenswrapper[4809]: I1205 11:11:12.059282 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:11:12 crc kubenswrapper[4809]: I1205 11:11:12.089229 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/700689a7-7e7b-473c-8f2d-cfdd43d8e8f9-metrics-certs\") pod \"network-metrics-daemon-xxkn4\" (UID: \"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9\") " pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:11:12 crc kubenswrapper[4809]: I1205 11:11:12.164823 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:12 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:12 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:12 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:12 crc kubenswrapper[4809]: I1205 11:11:12.164874 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:12 crc kubenswrapper[4809]: I1205 11:11:12.192363 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxkn4" Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.163222 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:13 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:13 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:13 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.163536 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.865954 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.866020 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.865965 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.866175 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.871773 4809 patch_prober.go:28] interesting pod/console-f9d7485db-zhtzr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 05 11:11:13 crc kubenswrapper[4809]: I1205 11:11:13.871827 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zhtzr" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 05 11:11:14 crc kubenswrapper[4809]: I1205 11:11:14.047045 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:11:14 crc kubenswrapper[4809]: I1205 11:11:14.047174 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:11:14 crc kubenswrapper[4809]: I1205 11:11:14.164611 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:14 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:14 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:14 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:14 crc kubenswrapper[4809]: I1205 11:11:14.164687 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:15 crc kubenswrapper[4809]: I1205 11:11:15.162524 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:15 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:15 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:15 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:15 crc kubenswrapper[4809]: I1205 11:11:15.163158 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:16 crc kubenswrapper[4809]: I1205 11:11:16.162035 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:16 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:16 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:16 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:16 crc kubenswrapper[4809]: I1205 11:11:16.162093 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:17 crc kubenswrapper[4809]: I1205 11:11:17.161810 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:17 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:17 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:17 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:17 crc kubenswrapper[4809]: I1205 11:11:17.161878 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:18 crc kubenswrapper[4809]: I1205 11:11:18.162801 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:18 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:18 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:18 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:18 crc kubenswrapper[4809]: I1205 11:11:18.163178 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:19 crc kubenswrapper[4809]: I1205 11:11:19.162143 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:19 crc kubenswrapper[4809]: [-]has-synced failed: reason withheld Dec 05 11:11:19 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:19 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:19 crc kubenswrapper[4809]: I1205 11:11:19.162225 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:20 crc kubenswrapper[4809]: I1205 11:11:20.162694 4809 patch_prober.go:28] interesting pod/router-default-5444994796-xgftq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 11:11:20 crc kubenswrapper[4809]: [+]has-synced ok Dec 05 11:11:20 crc kubenswrapper[4809]: [+]process-running ok Dec 05 11:11:20 crc kubenswrapper[4809]: healthz check failed Dec 05 11:11:20 crc kubenswrapper[4809]: I1205 11:11:20.163093 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xgftq" podUID="8d14e7df-f1f6-4a5d-964b-7f089e99d0fe" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 11:11:21 crc kubenswrapper[4809]: I1205 11:11:21.162672 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:11:21 crc kubenswrapper[4809]: I1205 11:11:21.164969 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-xgftq" Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.866059 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.866547 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.866668 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.867596 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"5a5dcf754013110a47bba8ad28287d43e7b8422c6b930367495bfd8465f881eb"} pod="openshift-console/downloads-7954f5f757-6fvxb" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.867818 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" containerID="cri-o://5a5dcf754013110a47bba8ad28287d43e7b8422c6b930367495bfd8465f881eb" gracePeriod=2 Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.866072 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.868053 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.868460 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.868525 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.872834 4809 patch_prober.go:28] interesting pod/console-f9d7485db-zhtzr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 05 11:11:23 crc kubenswrapper[4809]: I1205 11:11:23.872940 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zhtzr" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 05 11:11:24 crc kubenswrapper[4809]: I1205 11:11:24.559966 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:11:28 crc kubenswrapper[4809]: I1205 11:11:28.555730 4809 generic.go:334] "Generic (PLEG): container finished" podID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerID="5a5dcf754013110a47bba8ad28287d43e7b8422c6b930367495bfd8465f881eb" exitCode=0 Dec 05 11:11:28 crc kubenswrapper[4809]: I1205 11:11:28.555819 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6fvxb" event={"ID":"c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4","Type":"ContainerDied","Data":"5a5dcf754013110a47bba8ad28287d43e7b8422c6b930367495bfd8465f881eb"} Dec 05 11:11:32 crc kubenswrapper[4809]: I1205 11:11:32.706254 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 11:11:33 crc kubenswrapper[4809]: I1205 11:11:33.866790 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:33 crc kubenswrapper[4809]: I1205 11:11:33.866870 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:33 crc kubenswrapper[4809]: I1205 11:11:33.876216 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:11:33 crc kubenswrapper[4809]: I1205 11:11:33.879615 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:11:34 crc kubenswrapper[4809]: I1205 11:11:34.486336 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.329830 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.330871 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.335284 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.336862 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.342082 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.517308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.517365 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.618830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.619180 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.619235 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.642453 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:38 crc kubenswrapper[4809]: I1205 11:11:38.650374 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.519335 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.521314 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.532144 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.671196 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-var-lock\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.672049 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbdca384-2491-4744-913f-8be714d865a1-kube-api-access\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.672086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.773539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-var-lock\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.773857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbdca384-2491-4744-913f-8be714d865a1-kube-api-access\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.773980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.774084 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.773662 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-var-lock\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.792029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbdca384-2491-4744-913f-8be714d865a1-kube-api-access\") pod \"installer-9-crc\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:42 crc kubenswrapper[4809]: I1205 11:11:42.851387 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:11:43 crc kubenswrapper[4809]: E1205 11:11:43.415714 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:fcd9cdaeec4d21f010a2bb25043386ef71e3c6ca9c62aaf284b705dd309b1475: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:fcd9cdaeec4d21f010a2bb25043386ef71e3c6ca9c62aaf284b705dd309b1475\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 11:11:43 crc kubenswrapper[4809]: E1205 11:11:43.415934 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9hq4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qcp4f_openshift-marketplace(29be0f64-8b6e-453e-a8c0-dbe464da2705): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:fcd9cdaeec4d21f010a2bb25043386ef71e3c6ca9c62aaf284b705dd309b1475: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:fcd9cdaeec4d21f010a2bb25043386ef71e3c6ca9c62aaf284b705dd309b1475\": context canceled" logger="UnhandledError" Dec 05 11:11:43 crc kubenswrapper[4809]: E1205 11:11:43.417113 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:fcd9cdaeec4d21f010a2bb25043386ef71e3c6ca9c62aaf284b705dd309b1475: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:fcd9cdaeec4d21f010a2bb25043386ef71e3c6ca9c62aaf284b705dd309b1475\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-qcp4f" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" Dec 05 11:11:43 crc kubenswrapper[4809]: I1205 11:11:43.865316 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:43 crc kubenswrapper[4809]: I1205 11:11:43.865380 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:44 crc kubenswrapper[4809]: I1205 11:11:44.046819 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:11:44 crc kubenswrapper[4809]: I1205 11:11:44.046890 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:11:44 crc kubenswrapper[4809]: E1205 11:11:44.405239 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-qcp4f" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" Dec 05 11:11:45 crc kubenswrapper[4809]: E1205 11:11:45.548258 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 11:11:45 crc kubenswrapper[4809]: E1205 11:11:45.548686 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fkqqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-chdkw_openshift-marketplace(4bfa41b4-f65e-49a2-9b3c-783194dc934d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 11:11:45 crc kubenswrapper[4809]: E1205 11:11:45.550206 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-chdkw" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" Dec 05 11:11:49 crc kubenswrapper[4809]: E1205 11:11:49.882558 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 11:11:49 crc kubenswrapper[4809]: E1205 11:11:49.882746 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6m9b7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xj244_openshift-marketplace(5aaa8ff8-bbd1-4768-9325-285eb8acb01a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 11:11:49 crc kubenswrapper[4809]: E1205 11:11:49.884085 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xj244" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" Dec 05 11:11:50 crc kubenswrapper[4809]: E1205 11:11:50.713594 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xj244" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" Dec 05 11:11:50 crc kubenswrapper[4809]: E1205 11:11:50.713622 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-chdkw" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" Dec 05 11:11:53 crc kubenswrapper[4809]: I1205 11:11:53.865540 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:11:53 crc kubenswrapper[4809]: I1205 11:11:53.865935 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:11:55 crc kubenswrapper[4809]: E1205 11:11:55.775987 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 11:11:55 crc kubenswrapper[4809]: E1205 11:11:55.776431 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wqkrs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lfv24_openshift-marketplace(50cea908-3bdc-4a86-b577-47fa705d1957): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 11:11:55 crc kubenswrapper[4809]: E1205 11:11:55.777694 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lfv24" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" Dec 05 11:11:57 crc kubenswrapper[4809]: E1205 11:11:57.935479 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 11:11:57 crc kubenswrapper[4809]: E1205 11:11:57.935701 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nxjwx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9t5j4_openshift-marketplace(75df67eb-0996-4396-94bb-8ddbce4136ce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 11:11:57 crc kubenswrapper[4809]: E1205 11:11:57.936971 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9t5j4" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.293149 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lfv24" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.293218 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9t5j4" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.315887 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.316108 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gdfjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hlxqf_openshift-marketplace(b6f405ac-b705-477f-b150-109df1092b2b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.317701 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hlxqf" podUID="b6f405ac-b705-477f-b150-109df1092b2b" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.376547 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.377018 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrxvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jrg7v_openshift-marketplace(e93c2521-42a0-4944-97c7-448e498294a4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 11:12:00 crc kubenswrapper[4809]: E1205 11:12:00.378585 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jrg7v" podUID="e93c2521-42a0-4944-97c7-448e498294a4" Dec 05 11:12:03 crc kubenswrapper[4809]: I1205 11:12:03.865859 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:12:03 crc kubenswrapper[4809]: I1205 11:12:03.866291 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:12:10 crc kubenswrapper[4809]: E1205 11:12:10.213953 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jrg7v" podUID="e93c2521-42a0-4944-97c7-448e498294a4" Dec 05 11:12:10 crc kubenswrapper[4809]: E1205 11:12:10.214233 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-hlxqf" podUID="b6f405ac-b705-477f-b150-109df1092b2b" Dec 05 11:12:10 crc kubenswrapper[4809]: E1205 11:12:10.226903 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 11:12:10 crc kubenswrapper[4809]: E1205 11:12:10.227284 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4xbbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dmzbq_openshift-marketplace(73486c5f-7eeb-4a30-ad5c-2eaf64837d28): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 11:12:10 crc kubenswrapper[4809]: E1205 11:12:10.228432 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dmzbq" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.622378 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 11:12:10 crc kubenswrapper[4809]: W1205 11:12:10.644434 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbbdca384_2491_4744_913f_8be714d865a1.slice/crio-aa52cd9e6928a9b0fbff54be72fd932d14080b8df3e539e5df42a7e3903ddf65 WatchSource:0}: Error finding container aa52cd9e6928a9b0fbff54be72fd932d14080b8df3e539e5df42a7e3903ddf65: Status 404 returned error can't find the container with id aa52cd9e6928a9b0fbff54be72fd932d14080b8df3e539e5df42a7e3903ddf65 Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.669187 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xxkn4"] Dec 05 11:12:10 crc kubenswrapper[4809]: W1205 11:12:10.675969 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod700689a7_7e7b_473c_8f2d_cfdd43d8e8f9.slice/crio-a86bf232a1f47c16ff8d048108af051193a28f1b9c0a44d3d15e66d8c3592880 WatchSource:0}: Error finding container a86bf232a1f47c16ff8d048108af051193a28f1b9c0a44d3d15e66d8c3592880: Status 404 returned error can't find the container with id a86bf232a1f47c16ff8d048108af051193a28f1b9c0a44d3d15e66d8c3592880 Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.709305 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.782031 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" event={"ID":"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9","Type":"ContainerStarted","Data":"a86bf232a1f47c16ff8d048108af051193a28f1b9c0a44d3d15e66d8c3592880"} Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.783945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6fvxb" event={"ID":"c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4","Type":"ContainerStarted","Data":"00423d0b38782a1290731ce379f73f9d5e516bc220d325735dd8d27230c8483d"} Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.784823 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.784860 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.784826 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.785445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"072bb24a-5e0a-4caa-ae1d-997322c1dc3e","Type":"ContainerStarted","Data":"791fed230f8d2d06b76cae299318e197cba269978b18b687303ff2f499bde8b7"} Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.788502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bbdca384-2491-4744-913f-8be714d865a1","Type":"ContainerStarted","Data":"aa52cd9e6928a9b0fbff54be72fd932d14080b8df3e539e5df42a7e3903ddf65"} Dec 05 11:12:10 crc kubenswrapper[4809]: I1205 11:12:10.794244 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcp4f" event={"ID":"29be0f64-8b6e-453e-a8c0-dbe464da2705","Type":"ContainerStarted","Data":"20e630fd820bbc96a2e4f10a51fa3cc6705b8dc2bfddbf6e1e3023b24d322fb2"} Dec 05 11:12:10 crc kubenswrapper[4809]: E1205 11:12:10.815911 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dmzbq" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.802375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bbdca384-2491-4744-913f-8be714d865a1","Type":"ContainerStarted","Data":"7f7b8209c5d46c7f8e1f7e7cc463a67ec482e2f51678b54f43d53b2453356182"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.804298 4809 generic.go:334] "Generic (PLEG): container finished" podID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerID="20e630fd820bbc96a2e4f10a51fa3cc6705b8dc2bfddbf6e1e3023b24d322fb2" exitCode=0 Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.804391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcp4f" event={"ID":"29be0f64-8b6e-453e-a8c0-dbe464da2705","Type":"ContainerDied","Data":"20e630fd820bbc96a2e4f10a51fa3cc6705b8dc2bfddbf6e1e3023b24d322fb2"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.806622 4809 generic.go:334] "Generic (PLEG): container finished" podID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerID="f8d97c64c93f70d66aa091259bbd8c120436dbb7a847e37590fabc06667ce778" exitCode=0 Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.806656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9t5j4" event={"ID":"75df67eb-0996-4396-94bb-8ddbce4136ce","Type":"ContainerDied","Data":"f8d97c64c93f70d66aa091259bbd8c120436dbb7a847e37590fabc06667ce778"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.815272 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerID="4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742" exitCode=0 Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.815342 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chdkw" event={"ID":"4bfa41b4-f65e-49a2-9b3c-783194dc934d","Type":"ContainerDied","Data":"4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.821254 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=29.821238544 podStartE2EDuration="29.821238544s" podCreationTimestamp="2025-12-05 11:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:12:11.818678512 +0000 UTC m=+227.209655090" watchObservedRunningTime="2025-12-05 11:12:11.821238544 +0000 UTC m=+227.212215102" Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.829286 4809 generic.go:334] "Generic (PLEG): container finished" podID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerID="c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51" exitCode=0 Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.830285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xj244" event={"ID":"5aaa8ff8-bbd1-4768-9325-285eb8acb01a","Type":"ContainerDied","Data":"c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.842845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" event={"ID":"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9","Type":"ContainerStarted","Data":"71b5eca10cd043b64ea6088858fb93c552478f13af86143165f8d7f7fda1f6cf"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.842891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xxkn4" event={"ID":"700689a7-7e7b-473c-8f2d-cfdd43d8e8f9","Type":"ContainerStarted","Data":"9cbba8f80e3f7a816874674eef5e91364a126ab7b112abaece6a3b9763ca2039"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.846910 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"072bb24a-5e0a-4caa-ae1d-997322c1dc3e","Type":"ContainerStarted","Data":"fb7f459cd58b1541f2e1889ccbd1365d8f0a4937574c7e579645249b6d84fa30"} Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.847359 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.847411 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.910701 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=33.906608052 podStartE2EDuration="33.906608052s" podCreationTimestamp="2025-12-05 11:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:12:11.900047197 +0000 UTC m=+227.291023765" watchObservedRunningTime="2025-12-05 11:12:11.906608052 +0000 UTC m=+227.297584610" Dec 05 11:12:11 crc kubenswrapper[4809]: I1205 11:12:11.947283 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xxkn4" podStartSLOduration=203.947263169 podStartE2EDuration="3m23.947263169s" podCreationTimestamp="2025-12-05 11:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:12:11.926296238 +0000 UTC m=+227.317272826" watchObservedRunningTime="2025-12-05 11:12:11.947263169 +0000 UTC m=+227.338239727" Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.853359 4809 generic.go:334] "Generic (PLEG): container finished" podID="072bb24a-5e0a-4caa-ae1d-997322c1dc3e" containerID="fb7f459cd58b1541f2e1889ccbd1365d8f0a4937574c7e579645249b6d84fa30" exitCode=0 Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.853461 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"072bb24a-5e0a-4caa-ae1d-997322c1dc3e","Type":"ContainerDied","Data":"fb7f459cd58b1541f2e1889ccbd1365d8f0a4937574c7e579645249b6d84fa30"} Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.857095 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcp4f" event={"ID":"29be0f64-8b6e-453e-a8c0-dbe464da2705","Type":"ContainerStarted","Data":"4a4a8a556e52e0d31f19203c359c066f8865ba75e234a413712498d815eb0b00"} Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.859009 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9t5j4" event={"ID":"75df67eb-0996-4396-94bb-8ddbce4136ce","Type":"ContainerStarted","Data":"cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74"} Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.861250 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chdkw" event={"ID":"4bfa41b4-f65e-49a2-9b3c-783194dc934d","Type":"ContainerStarted","Data":"6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1"} Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.863463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xj244" event={"ID":"5aaa8ff8-bbd1-4768-9325-285eb8acb01a","Type":"ContainerStarted","Data":"74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de"} Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.864167 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.864225 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.911220 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qcp4f" podStartSLOduration=2.944213726 podStartE2EDuration="1m5.911194086s" podCreationTimestamp="2025-12-05 11:11:07 +0000 UTC" firstStartedPulling="2025-12-05 11:11:09.394884307 +0000 UTC m=+164.785860865" lastFinishedPulling="2025-12-05 11:12:12.361864667 +0000 UTC m=+227.752841225" observedRunningTime="2025-12-05 11:12:12.891037197 +0000 UTC m=+228.282013775" watchObservedRunningTime="2025-12-05 11:12:12.911194086 +0000 UTC m=+228.302170644" Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.911811 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9t5j4" podStartSLOduration=1.909676659 podStartE2EDuration="1m7.911804253s" podCreationTimestamp="2025-12-05 11:11:05 +0000 UTC" firstStartedPulling="2025-12-05 11:11:06.321642588 +0000 UTC m=+161.712619146" lastFinishedPulling="2025-12-05 11:12:12.323770182 +0000 UTC m=+227.714746740" observedRunningTime="2025-12-05 11:12:12.909515439 +0000 UTC m=+228.300491997" watchObservedRunningTime="2025-12-05 11:12:12.911804253 +0000 UTC m=+228.302780811" Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.929309 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xj244" podStartSLOduration=2.7559041520000003 podStartE2EDuration="1m7.929288756s" podCreationTimestamp="2025-12-05 11:11:05 +0000 UTC" firstStartedPulling="2025-12-05 11:11:07.332535327 +0000 UTC m=+162.723511885" lastFinishedPulling="2025-12-05 11:12:12.505919931 +0000 UTC m=+227.896896489" observedRunningTime="2025-12-05 11:12:12.928262898 +0000 UTC m=+228.319239456" watchObservedRunningTime="2025-12-05 11:12:12.929288756 +0000 UTC m=+228.320265314" Dec 05 11:12:12 crc kubenswrapper[4809]: I1205 11:12:12.951261 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-chdkw" podStartSLOduration=3.071662491 podStartE2EDuration="1m7.951235346s" podCreationTimestamp="2025-12-05 11:11:05 +0000 UTC" firstStartedPulling="2025-12-05 11:11:07.332522006 +0000 UTC m=+162.723498564" lastFinishedPulling="2025-12-05 11:12:12.212094861 +0000 UTC m=+227.603071419" observedRunningTime="2025-12-05 11:12:12.946742649 +0000 UTC m=+228.337719227" watchObservedRunningTime="2025-12-05 11:12:12.951235346 +0000 UTC m=+228.342211904" Dec 05 11:12:13 crc kubenswrapper[4809]: I1205 11:12:13.864772 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:12:13 crc kubenswrapper[4809]: I1205 11:12:13.864823 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:12:13 crc kubenswrapper[4809]: I1205 11:12:13.864905 4809 patch_prober.go:28] interesting pod/downloads-7954f5f757-6fvxb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 05 11:12:13 crc kubenswrapper[4809]: I1205 11:12:13.864928 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6fvxb" podUID="c8c80a2a-c020-4ad3-818d-e5b7bb52e8f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.046311 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.046360 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.046409 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.047040 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.047097 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4" gracePeriod=600 Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.175913 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.195190 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kube-api-access\") pod \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.195271 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kubelet-dir\") pod \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\" (UID: \"072bb24a-5e0a-4caa-ae1d-997322c1dc3e\") " Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.195402 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "072bb24a-5e0a-4caa-ae1d-997322c1dc3e" (UID: "072bb24a-5e0a-4caa-ae1d-997322c1dc3e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.200961 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "072bb24a-5e0a-4caa-ae1d-997322c1dc3e" (UID: "072bb24a-5e0a-4caa-ae1d-997322c1dc3e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.297102 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.297197 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/072bb24a-5e0a-4caa-ae1d-997322c1dc3e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.876600 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.880571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"072bb24a-5e0a-4caa-ae1d-997322c1dc3e","Type":"ContainerDied","Data":"791fed230f8d2d06b76cae299318e197cba269978b18b687303ff2f499bde8b7"} Dec 05 11:12:14 crc kubenswrapper[4809]: I1205 11:12:14.880608 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="791fed230f8d2d06b76cae299318e197cba269978b18b687303ff2f499bde8b7" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.348217 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.348276 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.568147 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.568217 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.884437 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4" exitCode=0 Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.884527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4"} Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.932144 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.934743 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.995489 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:12:15 crc kubenswrapper[4809]: I1205 11:12:15.995675 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:12:16 crc kubenswrapper[4809]: I1205 11:12:16.037790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:12:17 crc kubenswrapper[4809]: I1205 11:12:17.345070 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:12:17 crc kubenswrapper[4809]: I1205 11:12:17.345136 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:12:17 crc kubenswrapper[4809]: I1205 11:12:17.385314 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:12:17 crc kubenswrapper[4809]: I1205 11:12:17.899458 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"e7b326672db1ccbd336f3353e4dddf0e3ea107025106a857e28fa99fc3ca732a"} Dec 05 11:12:17 crc kubenswrapper[4809]: I1205 11:12:17.943967 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:12:17 crc kubenswrapper[4809]: I1205 11:12:17.963011 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:12:19 crc kubenswrapper[4809]: I1205 11:12:19.912624 4809 generic.go:334] "Generic (PLEG): container finished" podID="50cea908-3bdc-4a86-b577-47fa705d1957" containerID="f059a967b469b2185fd8103b609140ffe0183d070e09fe5002f639505b76d60e" exitCode=0 Dec 05 11:12:19 crc kubenswrapper[4809]: I1205 11:12:19.912787 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfv24" event={"ID":"50cea908-3bdc-4a86-b577-47fa705d1957","Type":"ContainerDied","Data":"f059a967b469b2185fd8103b609140ffe0183d070e09fe5002f639505b76d60e"} Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.313449 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-chdkw"] Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.313896 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-chdkw" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="registry-server" containerID="cri-o://6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1" gracePeriod=2 Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.692272 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.790671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-utilities\") pod \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.790717 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-catalog-content\") pod \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.790747 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkqqs\" (UniqueName: \"kubernetes.io/projected/4bfa41b4-f65e-49a2-9b3c-783194dc934d-kube-api-access-fkqqs\") pod \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\" (UID: \"4bfa41b4-f65e-49a2-9b3c-783194dc934d\") " Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.791777 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-utilities" (OuterVolumeSpecName: "utilities") pod "4bfa41b4-f65e-49a2-9b3c-783194dc934d" (UID: "4bfa41b4-f65e-49a2-9b3c-783194dc934d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.803797 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bfa41b4-f65e-49a2-9b3c-783194dc934d-kube-api-access-fkqqs" (OuterVolumeSpecName: "kube-api-access-fkqqs") pod "4bfa41b4-f65e-49a2-9b3c-783194dc934d" (UID: "4bfa41b4-f65e-49a2-9b3c-783194dc934d"). InnerVolumeSpecName "kube-api-access-fkqqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.860257 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bfa41b4-f65e-49a2-9b3c-783194dc934d" (UID: "4bfa41b4-f65e-49a2-9b3c-783194dc934d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.893026 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.893072 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bfa41b4-f65e-49a2-9b3c-783194dc934d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.893086 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkqqs\" (UniqueName: \"kubernetes.io/projected/4bfa41b4-f65e-49a2-9b3c-783194dc934d-kube-api-access-fkqqs\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.925975 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfv24" event={"ID":"50cea908-3bdc-4a86-b577-47fa705d1957","Type":"ContainerStarted","Data":"8173b79fa841b55fae16ad578702de25ccf43ee4653f48a736c64df08bf35647"} Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.928553 4809 generic.go:334] "Generic (PLEG): container finished" podID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerID="6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1" exitCode=0 Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.928581 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chdkw" event={"ID":"4bfa41b4-f65e-49a2-9b3c-783194dc934d","Type":"ContainerDied","Data":"6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1"} Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.928614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-chdkw" event={"ID":"4bfa41b4-f65e-49a2-9b3c-783194dc934d","Type":"ContainerDied","Data":"e39306b9f3d349ec761e00f33607cbedfbc07b459df76f2f855a6f33f9bbee73"} Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.928659 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-chdkw" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.928662 4809 scope.go:117] "RemoveContainer" containerID="6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.949376 4809 scope.go:117] "RemoveContainer" containerID="4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.952206 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lfv24" podStartSLOduration=2.855812721 podStartE2EDuration="1m15.952190033s" podCreationTimestamp="2025-12-05 11:11:05 +0000 UTC" firstStartedPulling="2025-12-05 11:11:07.346888882 +0000 UTC m=+162.737865450" lastFinishedPulling="2025-12-05 11:12:20.443266194 +0000 UTC m=+235.834242762" observedRunningTime="2025-12-05 11:12:20.948947862 +0000 UTC m=+236.339924420" watchObservedRunningTime="2025-12-05 11:12:20.952190033 +0000 UTC m=+236.343166591" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.964284 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-chdkw"] Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.970563 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-chdkw"] Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.973812 4809 scope.go:117] "RemoveContainer" containerID="2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.991327 4809 scope.go:117] "RemoveContainer" containerID="6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1" Dec 05 11:12:20 crc kubenswrapper[4809]: E1205 11:12:20.991892 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1\": container with ID starting with 6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1 not found: ID does not exist" containerID="6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.991921 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1"} err="failed to get container status \"6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1\": rpc error: code = NotFound desc = could not find container \"6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1\": container with ID starting with 6d179917cab9b16cd53999b0d1167354896f20745317085010fae5c88f983bd1 not found: ID does not exist" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.991941 4809 scope.go:117] "RemoveContainer" containerID="4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742" Dec 05 11:12:20 crc kubenswrapper[4809]: E1205 11:12:20.992503 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742\": container with ID starting with 4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742 not found: ID does not exist" containerID="4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.992564 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742"} err="failed to get container status \"4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742\": rpc error: code = NotFound desc = could not find container \"4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742\": container with ID starting with 4a0606b82fda990c653860fc669b40d1ec511e70cc709232a3f8e2ed29c6a742 not found: ID does not exist" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.992618 4809 scope.go:117] "RemoveContainer" containerID="2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993" Dec 05 11:12:20 crc kubenswrapper[4809]: E1205 11:12:20.993093 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993\": container with ID starting with 2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993 not found: ID does not exist" containerID="2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993" Dec 05 11:12:20 crc kubenswrapper[4809]: I1205 11:12:20.993133 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993"} err="failed to get container status \"2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993\": rpc error: code = NotFound desc = could not find container \"2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993\": container with ID starting with 2ef8f1b2f138fdd6179e6869a53d033cbaf862184250d0143da5d6bb4ac4d993 not found: ID does not exist" Dec 05 11:12:22 crc kubenswrapper[4809]: I1205 11:12:22.883372 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" path="/var/lib/kubelet/pods/4bfa41b4-f65e-49a2-9b3c-783194dc934d/volumes" Dec 05 11:12:22 crc kubenswrapper[4809]: I1205 11:12:22.940963 4809 generic.go:334] "Generic (PLEG): container finished" podID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerID="e27c5b23e8c1fa7f6a1e06074226f0cb2e6f8f2d1b2660dccd607016d0ca614d" exitCode=0 Dec 05 11:12:22 crc kubenswrapper[4809]: I1205 11:12:22.941008 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmzbq" event={"ID":"73486c5f-7eeb-4a30-ad5c-2eaf64837d28","Type":"ContainerDied","Data":"e27c5b23e8c1fa7f6a1e06074226f0cb2e6f8f2d1b2660dccd607016d0ca614d"} Dec 05 11:12:23 crc kubenswrapper[4809]: I1205 11:12:23.882022 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6fvxb" Dec 05 11:12:25 crc kubenswrapper[4809]: I1205 11:12:25.392516 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:12:25 crc kubenswrapper[4809]: I1205 11:12:25.605710 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:12:25 crc kubenswrapper[4809]: I1205 11:12:25.751497 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:12:25 crc kubenswrapper[4809]: I1205 11:12:25.751560 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:12:25 crc kubenswrapper[4809]: I1205 11:12:25.798843 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:12:26 crc kubenswrapper[4809]: I1205 11:12:26.007973 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:12:28 crc kubenswrapper[4809]: I1205 11:12:28.111152 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lfv24"] Dec 05 11:12:28 crc kubenswrapper[4809]: I1205 11:12:28.112328 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lfv24" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="registry-server" containerID="cri-o://8173b79fa841b55fae16ad578702de25ccf43ee4653f48a736c64df08bf35647" gracePeriod=2 Dec 05 11:12:31 crc kubenswrapper[4809]: I1205 11:12:31.984411 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lfv24_50cea908-3bdc-4a86-b577-47fa705d1957/registry-server/0.log" Dec 05 11:12:31 crc kubenswrapper[4809]: I1205 11:12:31.985480 4809 generic.go:334] "Generic (PLEG): container finished" podID="50cea908-3bdc-4a86-b577-47fa705d1957" containerID="8173b79fa841b55fae16ad578702de25ccf43ee4653f48a736c64df08bf35647" exitCode=137 Dec 05 11:12:31 crc kubenswrapper[4809]: I1205 11:12:31.985508 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfv24" event={"ID":"50cea908-3bdc-4a86-b577-47fa705d1957","Type":"ContainerDied","Data":"8173b79fa841b55fae16ad578702de25ccf43ee4653f48a736c64df08bf35647"} Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.488602 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lfv24_50cea908-3bdc-4a86-b577-47fa705d1957/registry-server/0.log" Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.489779 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.689116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-utilities\") pod \"50cea908-3bdc-4a86-b577-47fa705d1957\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.689348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqkrs\" (UniqueName: \"kubernetes.io/projected/50cea908-3bdc-4a86-b577-47fa705d1957-kube-api-access-wqkrs\") pod \"50cea908-3bdc-4a86-b577-47fa705d1957\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.689576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-catalog-content\") pod \"50cea908-3bdc-4a86-b577-47fa705d1957\" (UID: \"50cea908-3bdc-4a86-b577-47fa705d1957\") " Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.691521 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-utilities" (OuterVolumeSpecName: "utilities") pod "50cea908-3bdc-4a86-b577-47fa705d1957" (UID: "50cea908-3bdc-4a86-b577-47fa705d1957"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.698031 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50cea908-3bdc-4a86-b577-47fa705d1957-kube-api-access-wqkrs" (OuterVolumeSpecName: "kube-api-access-wqkrs") pod "50cea908-3bdc-4a86-b577-47fa705d1957" (UID: "50cea908-3bdc-4a86-b577-47fa705d1957"). InnerVolumeSpecName "kube-api-access-wqkrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.739486 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50cea908-3bdc-4a86-b577-47fa705d1957" (UID: "50cea908-3bdc-4a86-b577-47fa705d1957"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.790361 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqkrs\" (UniqueName: \"kubernetes.io/projected/50cea908-3bdc-4a86-b577-47fa705d1957-kube-api-access-wqkrs\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.790399 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:32 crc kubenswrapper[4809]: I1205 11:12:32.790412 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50cea908-3bdc-4a86-b577-47fa705d1957-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.002342 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lfv24_50cea908-3bdc-4a86-b577-47fa705d1957/registry-server/0.log" Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.005939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lfv24" event={"ID":"50cea908-3bdc-4a86-b577-47fa705d1957","Type":"ContainerDied","Data":"afea6276daff65f99c186ef5de375ca6b39325febd4b18885e34d3094f90a3f9"} Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.005999 4809 scope.go:117] "RemoveContainer" containerID="8173b79fa841b55fae16ad578702de25ccf43ee4653f48a736c64df08bf35647" Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.006936 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lfv24" Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.010852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxqf" event={"ID":"b6f405ac-b705-477f-b150-109df1092b2b","Type":"ContainerStarted","Data":"99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8"} Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.014085 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmzbq" event={"ID":"73486c5f-7eeb-4a30-ad5c-2eaf64837d28","Type":"ContainerStarted","Data":"3ae61dc78d40469c6125bbcc4622c5dc6de098b77c497861f1f22062bb839eff"} Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.017485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrg7v" event={"ID":"e93c2521-42a0-4944-97c7-448e498294a4","Type":"ContainerStarted","Data":"769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461"} Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.025519 4809 scope.go:117] "RemoveContainer" containerID="f059a967b469b2185fd8103b609140ffe0183d070e09fe5002f639505b76d60e" Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.047491 4809 scope.go:117] "RemoveContainer" containerID="494545a5aca2d31affd87654fa17184d9898bfec1b700ccff272628697713457" Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.051706 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lfv24"] Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.054184 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lfv24"] Dec 05 11:12:33 crc kubenswrapper[4809]: I1205 11:12:33.081700 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dmzbq" podStartSLOduration=3.987244251 podStartE2EDuration="1m26.08168405s" podCreationTimestamp="2025-12-05 11:11:07 +0000 UTC" firstStartedPulling="2025-12-05 11:11:10.434683361 +0000 UTC m=+165.825659919" lastFinishedPulling="2025-12-05 11:12:32.52912316 +0000 UTC m=+247.920099718" observedRunningTime="2025-12-05 11:12:33.079294523 +0000 UTC m=+248.470271081" watchObservedRunningTime="2025-12-05 11:12:33.08168405 +0000 UTC m=+248.472660608" Dec 05 11:12:34 crc kubenswrapper[4809]: I1205 11:12:34.024789 4809 generic.go:334] "Generic (PLEG): container finished" podID="e93c2521-42a0-4944-97c7-448e498294a4" containerID="769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461" exitCode=0 Dec 05 11:12:34 crc kubenswrapper[4809]: I1205 11:12:34.024843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrg7v" event={"ID":"e93c2521-42a0-4944-97c7-448e498294a4","Type":"ContainerDied","Data":"769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461"} Dec 05 11:12:34 crc kubenswrapper[4809]: I1205 11:12:34.030876 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6f405ac-b705-477f-b150-109df1092b2b" containerID="99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8" exitCode=0 Dec 05 11:12:34 crc kubenswrapper[4809]: I1205 11:12:34.030921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxqf" event={"ID":"b6f405ac-b705-477f-b150-109df1092b2b","Type":"ContainerDied","Data":"99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8"} Dec 05 11:12:34 crc kubenswrapper[4809]: I1205 11:12:34.572823 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-t4hqr"] Dec 05 11:12:34 crc kubenswrapper[4809]: I1205 11:12:34.878317 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" path="/var/lib/kubelet/pods/50cea908-3bdc-4a86-b577-47fa705d1957/volumes" Dec 05 11:12:35 crc kubenswrapper[4809]: I1205 11:12:35.037209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxqf" event={"ID":"b6f405ac-b705-477f-b150-109df1092b2b","Type":"ContainerStarted","Data":"4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b"} Dec 05 11:12:35 crc kubenswrapper[4809]: I1205 11:12:35.053518 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hlxqf" podStartSLOduration=2.831532166 podStartE2EDuration="1m27.053498912s" podCreationTimestamp="2025-12-05 11:11:08 +0000 UTC" firstStartedPulling="2025-12-05 11:11:10.442334127 +0000 UTC m=+165.833310685" lastFinishedPulling="2025-12-05 11:12:34.664300873 +0000 UTC m=+250.055277431" observedRunningTime="2025-12-05 11:12:35.051331811 +0000 UTC m=+250.442308369" watchObservedRunningTime="2025-12-05 11:12:35.053498912 +0000 UTC m=+250.444475470" Dec 05 11:12:36 crc kubenswrapper[4809]: I1205 11:12:36.043145 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrg7v" event={"ID":"e93c2521-42a0-4944-97c7-448e498294a4","Type":"ContainerStarted","Data":"1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92"} Dec 05 11:12:36 crc kubenswrapper[4809]: I1205 11:12:36.062174 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jrg7v" podStartSLOduration=3.595096202 podStartE2EDuration="1m28.062154791s" podCreationTimestamp="2025-12-05 11:11:08 +0000 UTC" firstStartedPulling="2025-12-05 11:11:10.440260509 +0000 UTC m=+165.831237067" lastFinishedPulling="2025-12-05 11:12:34.907318608 +0000 UTC m=+250.298295656" observedRunningTime="2025-12-05 11:12:36.061491192 +0000 UTC m=+251.452467770" watchObservedRunningTime="2025-12-05 11:12:36.062154791 +0000 UTC m=+251.453131349" Dec 05 11:12:37 crc kubenswrapper[4809]: I1205 11:12:37.866808 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:12:37 crc kubenswrapper[4809]: I1205 11:12:37.867198 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:12:38 crc kubenswrapper[4809]: I1205 11:12:38.774587 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:12:38 crc kubenswrapper[4809]: I1205 11:12:38.774666 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:12:38 crc kubenswrapper[4809]: I1205 11:12:38.904922 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-dmzbq" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="registry-server" probeResult="failure" output=< Dec 05 11:12:38 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 11:12:38 crc kubenswrapper[4809]: > Dec 05 11:12:39 crc kubenswrapper[4809]: I1205 11:12:39.168305 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:12:39 crc kubenswrapper[4809]: I1205 11:12:39.168406 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:12:39 crc kubenswrapper[4809]: I1205 11:12:39.816850 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jrg7v" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="registry-server" probeResult="failure" output=< Dec 05 11:12:39 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 11:12:39 crc kubenswrapper[4809]: > Dec 05 11:12:40 crc kubenswrapper[4809]: I1205 11:12:40.207797 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hlxqf" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="registry-server" probeResult="failure" output=< Dec 05 11:12:40 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 11:12:40 crc kubenswrapper[4809]: > Dec 05 11:12:47 crc kubenswrapper[4809]: I1205 11:12:47.914863 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:12:47 crc kubenswrapper[4809]: I1205 11:12:47.955454 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:12:48 crc kubenswrapper[4809]: I1205 11:12:48.142524 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmzbq"] Dec 05 11:12:48 crc kubenswrapper[4809]: I1205 11:12:48.821753 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:12:48 crc kubenswrapper[4809]: I1205 11:12:48.868898 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:12:49 crc kubenswrapper[4809]: I1205 11:12:49.111280 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dmzbq" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="registry-server" containerID="cri-o://3ae61dc78d40469c6125bbcc4622c5dc6de098b77c497861f1f22062bb839eff" gracePeriod=2 Dec 05 11:12:49 crc kubenswrapper[4809]: I1205 11:12:49.216732 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:12:49 crc kubenswrapper[4809]: I1205 11:12:49.257910 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:12:50 crc kubenswrapper[4809]: I1205 11:12:50.541338 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hlxqf"] Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.122851 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hlxqf" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="registry-server" containerID="cri-o://4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b" gracePeriod=2 Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977478 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.977824 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="extract-utilities" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977837 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="extract-utilities" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.977851 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="extract-content" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977858 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="extract-content" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.977871 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="registry-server" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977878 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="registry-server" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.977893 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="extract-content" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977899 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="extract-content" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.977907 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="extract-utilities" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977913 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="extract-utilities" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.977925 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="registry-server" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977931 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="registry-server" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.977943 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072bb24a-5e0a-4caa-ae1d-997322c1dc3e" containerName="pruner" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.977950 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="072bb24a-5e0a-4caa-ae1d-997322c1dc3e" containerName="pruner" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978053 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bfa41b4-f65e-49a2-9b3c-783194dc934d" containerName="registry-server" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978069 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50cea908-3bdc-4a86-b577-47fa705d1957" containerName="registry-server" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978080 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="072bb24a-5e0a-4caa-ae1d-997322c1dc3e" containerName="pruner" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978461 4809 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978602 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978758 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d" gracePeriod=15 Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978768 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510" gracePeriod=15 Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978816 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d" gracePeriod=15 Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978896 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a" gracePeriod=15 Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.978878 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa" gracePeriod=15 Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.979092 4809 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.979587 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.979611 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.979717 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.979730 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.979743 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.979754 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.979771 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.979783 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.979797 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.979806 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 11:12:51 crc kubenswrapper[4809]: E1205 11:12:51.979817 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.979825 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.980015 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.980032 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.980047 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.980060 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 11:12:51 crc kubenswrapper[4809]: I1205 11:12:51.980069 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.021534 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.143692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.143751 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.143779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.143802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.143819 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.143866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.143999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.144085 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245514 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245645 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245713 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245717 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245799 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245820 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245913 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245924 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.245991 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: I1205 11:12:52.319197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:12:52 crc kubenswrapper[4809]: W1205 11:12:52.366719 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-781e2a19fdb4b868f51724f10dbd93cbc196154ec42986eb190eaf363f1b72c7 WatchSource:0}: Error finding container 781e2a19fdb4b868f51724f10dbd93cbc196154ec42986eb190eaf363f1b72c7: Status 404 returned error can't find the container with id 781e2a19fdb4b868f51724f10dbd93cbc196154ec42986eb190eaf363f1b72c7 Dec 05 11:12:52 crc kubenswrapper[4809]: E1205 11:12:52.370274 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e4d5e875dd255 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 11:12:52.369150549 +0000 UTC m=+267.760127107,LastTimestamp:2025-12-05 11:12:52.369150549 +0000 UTC m=+267.760127107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.136691 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.137775 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510" exitCode=0 Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.137811 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa" exitCode=0 Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.137823 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d" exitCode=0 Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.137831 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a" exitCode=2 Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.139596 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dmzbq_73486c5f-7eeb-4a30-ad5c-2eaf64837d28/registry-server/0.log" Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.140424 4809 generic.go:334] "Generic (PLEG): container finished" podID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerID="3ae61dc78d40469c6125bbcc4622c5dc6de098b77c497861f1f22062bb839eff" exitCode=137 Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.140453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmzbq" event={"ID":"73486c5f-7eeb-4a30-ad5c-2eaf64837d28","Type":"ContainerDied","Data":"3ae61dc78d40469c6125bbcc4622c5dc6de098b77c497861f1f22062bb839eff"} Dec 05 11:12:53 crc kubenswrapper[4809]: I1205 11:12:53.141428 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"781e2a19fdb4b868f51724f10dbd93cbc196154ec42986eb190eaf363f1b72c7"} Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.875049 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.940121 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dmzbq_73486c5f-7eeb-4a30-ad5c-2eaf64837d28/registry-server/0.log" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.941030 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.941438 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.941911 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.947149 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.948294 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.948710 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.949036 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.949393 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980365 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980424 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-catalog-content\") pod \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-utilities\") pod \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980487 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980545 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xbbf\" (UniqueName: \"kubernetes.io/projected/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-kube-api-access-4xbbf\") pod \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\" (UID: \"73486c5f-7eeb-4a30-ad5c-2eaf64837d28\") " Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.980831 4809 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.981253 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.981252 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.981435 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-utilities" (OuterVolumeSpecName: "utilities") pod "73486c5f-7eeb-4a30-ad5c-2eaf64837d28" (UID: "73486c5f-7eeb-4a30-ad5c-2eaf64837d28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:54 crc kubenswrapper[4809]: I1205 11:12:54.987839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-kube-api-access-4xbbf" (OuterVolumeSpecName: "kube-api-access-4xbbf") pod "73486c5f-7eeb-4a30-ad5c-2eaf64837d28" (UID: "73486c5f-7eeb-4a30-ad5c-2eaf64837d28"). InnerVolumeSpecName "kube-api-access-4xbbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.011724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73486c5f-7eeb-4a30-ad5c-2eaf64837d28" (UID: "73486c5f-7eeb-4a30-ad5c-2eaf64837d28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.081455 4809 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.081491 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.081500 4809 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.081510 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xbbf\" (UniqueName: \"kubernetes.io/projected/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-kube-api-access-4xbbf\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.081519 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73486c5f-7eeb-4a30-ad5c-2eaf64837d28-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.135052 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.135461 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.135790 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.136305 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.136944 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.151882 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe"} Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.152447 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.152669 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.152889 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.153059 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.153837 4809 generic.go:334] "Generic (PLEG): container finished" podID="bbdca384-2491-4744-913f-8be714d865a1" containerID="7f7b8209c5d46c7f8e1f7e7cc463a67ec482e2f51678b54f43d53b2453356182" exitCode=0 Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.153921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bbdca384-2491-4744-913f-8be714d865a1","Type":"ContainerDied","Data":"7f7b8209c5d46c7f8e1f7e7cc463a67ec482e2f51678b54f43d53b2453356182"} Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.154380 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.154623 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.154851 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.155153 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.155363 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.157411 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.158209 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d" exitCode=0 Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.158308 4809 scope.go:117] "RemoveContainer" containerID="4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.158432 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.160138 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dmzbq_73486c5f-7eeb-4a30-ad5c-2eaf64837d28/registry-server/0.log" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.163037 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmzbq" event={"ID":"73486c5f-7eeb-4a30-ad5c-2eaf64837d28","Type":"ContainerDied","Data":"8ba555b4d9ceff40801fd9fdad25448b1e55c2b567ea41cd22ec0b468e547785"} Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.163113 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmzbq" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.164421 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.164852 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.165394 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.165717 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.166622 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.168666 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6f405ac-b705-477f-b150-109df1092b2b" containerID="4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b" exitCode=0 Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.168700 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxqf" event={"ID":"b6f405ac-b705-477f-b150-109df1092b2b","Type":"ContainerDied","Data":"4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b"} Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.168803 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlxqf" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.168734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlxqf" event={"ID":"b6f405ac-b705-477f-b150-109df1092b2b","Type":"ContainerDied","Data":"b9545c67a2ad6672c39dbc6bb8267b2f83f9d6c5f09fea92a1c19bce5f159e83"} Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.169422 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.170038 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.170299 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.170526 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.170781 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.177796 4809 scope.go:117] "RemoveContainer" containerID="46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.181121 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.181434 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.181680 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.181975 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.182303 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.182771 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.182941 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.183083 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.183219 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.183355 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.192500 4809 scope.go:117] "RemoveContainer" containerID="f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.208698 4809 scope.go:117] "RemoveContainer" containerID="ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.223156 4809 scope.go:117] "RemoveContainer" containerID="98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.240390 4809 scope.go:117] "RemoveContainer" containerID="a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.260904 4809 scope.go:117] "RemoveContainer" containerID="4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.262042 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\": container with ID starting with 4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510 not found: ID does not exist" containerID="4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.262091 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510"} err="failed to get container status \"4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\": rpc error: code = NotFound desc = could not find container \"4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510\": container with ID starting with 4facf6ece76e9834b215991ee78a31ec06462f6f6969d03e5b7ad13c76243510 not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.262123 4809 scope.go:117] "RemoveContainer" containerID="46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.262522 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\": container with ID starting with 46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa not found: ID does not exist" containerID="46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.262545 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa"} err="failed to get container status \"46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\": rpc error: code = NotFound desc = could not find container \"46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa\": container with ID starting with 46d39d8916315ca79d9ad1872d3413a623441c2b6498ac56ea5d22a8fcb8d0fa not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.262558 4809 scope.go:117] "RemoveContainer" containerID="f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.263741 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\": container with ID starting with f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d not found: ID does not exist" containerID="f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.263763 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d"} err="failed to get container status \"f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\": rpc error: code = NotFound desc = could not find container \"f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d\": container with ID starting with f4ae179e640476228e6f1709154e8f967e71efc968673162051e76f88e5b552d not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.263778 4809 scope.go:117] "RemoveContainer" containerID="ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.264196 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\": container with ID starting with ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a not found: ID does not exist" containerID="ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.264289 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a"} err="failed to get container status \"ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\": rpc error: code = NotFound desc = could not find container \"ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a\": container with ID starting with ad8569d83b37606f1977a9e2618bf7abe4a43140d49e7dd894b003a57da1b41a not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.264337 4809 scope.go:117] "RemoveContainer" containerID="98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.265855 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\": container with ID starting with 98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d not found: ID does not exist" containerID="98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.265896 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d"} err="failed to get container status \"98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\": rpc error: code = NotFound desc = could not find container \"98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d\": container with ID starting with 98efba8e3d086c3791b6112399d5f7a9229af4f18136007e9e25c085e165126d not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.265923 4809 scope.go:117] "RemoveContainer" containerID="a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.266277 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\": container with ID starting with a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8 not found: ID does not exist" containerID="a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.266355 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8"} err="failed to get container status \"a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\": rpc error: code = NotFound desc = could not find container \"a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8\": container with ID starting with a1c2cd73b46c8c6d935707a54f070970f38a5e39a949d2237ed8ae0317d52ab8 not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.266376 4809 scope.go:117] "RemoveContainer" containerID="3ae61dc78d40469c6125bbcc4622c5dc6de098b77c497861f1f22062bb839eff" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.280320 4809 scope.go:117] "RemoveContainer" containerID="e27c5b23e8c1fa7f6a1e06074226f0cb2e6f8f2d1b2660dccd607016d0ca614d" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.283485 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-catalog-content\") pod \"b6f405ac-b705-477f-b150-109df1092b2b\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.283533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/b6f405ac-b705-477f-b150-109df1092b2b-kube-api-access-gdfjz\") pod \"b6f405ac-b705-477f-b150-109df1092b2b\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.283599 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-utilities\") pod \"b6f405ac-b705-477f-b150-109df1092b2b\" (UID: \"b6f405ac-b705-477f-b150-109df1092b2b\") " Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.284753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-utilities" (OuterVolumeSpecName: "utilities") pod "b6f405ac-b705-477f-b150-109df1092b2b" (UID: "b6f405ac-b705-477f-b150-109df1092b2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.288694 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f405ac-b705-477f-b150-109df1092b2b-kube-api-access-gdfjz" (OuterVolumeSpecName: "kube-api-access-gdfjz") pod "b6f405ac-b705-477f-b150-109df1092b2b" (UID: "b6f405ac-b705-477f-b150-109df1092b2b"). InnerVolumeSpecName "kube-api-access-gdfjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.296079 4809 scope.go:117] "RemoveContainer" containerID="97eeb0e7d0a10e20a743d7895ae6271578c4ef011ec33af5fc054c54eadce737" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.318870 4809 scope.go:117] "RemoveContainer" containerID="4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.335390 4809 scope.go:117] "RemoveContainer" containerID="99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.355511 4809 scope.go:117] "RemoveContainer" containerID="bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.385174 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/b6f405ac-b705-477f-b150-109df1092b2b-kube-api-access-gdfjz\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.385225 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.387108 4809 scope.go:117] "RemoveContainer" containerID="4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.388291 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b\": container with ID starting with 4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b not found: ID does not exist" containerID="4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.388334 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b"} err="failed to get container status \"4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b\": rpc error: code = NotFound desc = could not find container \"4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b\": container with ID starting with 4cc33a640d1797fca6bc7ec02f633002b667f8e1af7453092c8fd463bb90798b not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.388372 4809 scope.go:117] "RemoveContainer" containerID="99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.388941 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8\": container with ID starting with 99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8 not found: ID does not exist" containerID="99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.388990 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8"} err="failed to get container status \"99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8\": rpc error: code = NotFound desc = could not find container \"99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8\": container with ID starting with 99bda55b7a0a69dab81ba5113026fb1d8cfaa6ee9abbb28eb89914dfc544caf8 not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.389063 4809 scope.go:117] "RemoveContainer" containerID="bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca" Dec 05 11:12:55 crc kubenswrapper[4809]: E1205 11:12:55.389544 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca\": container with ID starting with bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca not found: ID does not exist" containerID="bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.389574 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca"} err="failed to get container status \"bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca\": rpc error: code = NotFound desc = could not find container \"bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca\": container with ID starting with bc8bb563e71c161ef453706c8725cf1c328b6c6e247cbfa622a8a62ba76725ca not found: ID does not exist" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.402584 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6f405ac-b705-477f-b150-109df1092b2b" (UID: "b6f405ac-b705-477f-b150-109df1092b2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.485477 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.485761 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.485974 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.486145 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.486377 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:55 crc kubenswrapper[4809]: I1205 11:12:55.486962 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6f405ac-b705-477f-b150-109df1092b2b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.443254 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.444040 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.444363 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.444780 4809 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.444983 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.445208 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.601318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbdca384-2491-4744-913f-8be714d865a1-kube-api-access\") pod \"bbdca384-2491-4744-913f-8be714d865a1\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.601392 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-kubelet-dir\") pod \"bbdca384-2491-4744-913f-8be714d865a1\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.601468 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-var-lock\") pod \"bbdca384-2491-4744-913f-8be714d865a1\" (UID: \"bbdca384-2491-4744-913f-8be714d865a1\") " Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.601810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-var-lock" (OuterVolumeSpecName: "var-lock") pod "bbdca384-2491-4744-913f-8be714d865a1" (UID: "bbdca384-2491-4744-913f-8be714d865a1"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.601850 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bbdca384-2491-4744-913f-8be714d865a1" (UID: "bbdca384-2491-4744-913f-8be714d865a1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.606467 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbdca384-2491-4744-913f-8be714d865a1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bbdca384-2491-4744-913f-8be714d865a1" (UID: "bbdca384-2491-4744-913f-8be714d865a1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.702477 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbdca384-2491-4744-913f-8be714d865a1-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.702509 4809 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.702521 4809 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bbdca384-2491-4744-913f-8be714d865a1-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 11:12:56 crc kubenswrapper[4809]: I1205 11:12:56.880283 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 05 11:12:57 crc kubenswrapper[4809]: I1205 11:12:57.189117 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bbdca384-2491-4744-913f-8be714d865a1","Type":"ContainerDied","Data":"aa52cd9e6928a9b0fbff54be72fd932d14080b8df3e539e5df42a7e3903ddf65"} Dec 05 11:12:57 crc kubenswrapper[4809]: I1205 11:12:57.189168 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa52cd9e6928a9b0fbff54be72fd932d14080b8df3e539e5df42a7e3903ddf65" Dec 05 11:12:57 crc kubenswrapper[4809]: I1205 11:12:57.189187 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 11:12:57 crc kubenswrapper[4809]: I1205 11:12:57.194486 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:57 crc kubenswrapper[4809]: I1205 11:12:57.195677 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:57 crc kubenswrapper[4809]: I1205 11:12:57.196093 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:57 crc kubenswrapper[4809]: I1205 11:12:57.196442 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.176493 4809 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e4d5e875dd255 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 11:12:52.369150549 +0000 UTC m=+267.760127107,LastTimestamp:2025-12-05 11:12:52.369150549 +0000 UTC m=+267.760127107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.236151 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.236867 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.238274 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.238590 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.238840 4809 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.238866 4809 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.239015 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.440456 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.599574 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" containerName="oauth-openshift" containerID="cri-o://8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678" gracePeriod=15 Dec 05 11:12:59 crc kubenswrapper[4809]: E1205 11:12:59.841892 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.941809 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.942502 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.942903 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.943191 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.943487 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:12:59 crc kubenswrapper[4809]: I1205 11:12:59.943756 4809 status_manager.go:851] "Failed to get status for pod" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4hqr\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.140866 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6p4v\" (UniqueName: \"kubernetes.io/projected/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-kube-api-access-z6p4v\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.140932 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-dir\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.140961 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-error\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.141009 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-service-ca\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.141032 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-ocp-branding-template\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.141097 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.141829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.142432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-login\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.142589 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-serving-cert\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.142923 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-cliconfig\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.143094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-provider-selection\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.143256 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-idp-0-file-data\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.143419 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-trusted-ca-bundle\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.144322 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-router-certs\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.144495 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-session\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.144706 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-policies\") pod \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\" (UID: \"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b\") " Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.145202 4809 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.145915 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.143325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.144246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.145978 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.149117 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.154321 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-kube-api-access-z6p4v" (OuterVolumeSpecName: "kube-api-access-z6p4v") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "kube-api-access-z6p4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.154567 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.154973 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.155395 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.155469 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.155767 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.155918 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.156130 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" (UID: "1ac8b761-cb00-4b83-bc1e-0da82ee5c88b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.224376 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" containerID="8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678" exitCode=0 Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.224431 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" event={"ID":"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b","Type":"ContainerDied","Data":"8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678"} Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.224463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" event={"ID":"1ac8b761-cb00-4b83-bc1e-0da82ee5c88b","Type":"ContainerDied","Data":"d9201b743f1677e7e244b5af02f92edb08995cff5ec6b34dea455e6a9ec2e7a2"} Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.224482 4809 scope.go:117] "RemoveContainer" containerID="8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.224599 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.227983 4809 status_manager.go:851] "Failed to get status for pod" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4hqr\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.228570 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.228829 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.229455 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.229836 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.243840 4809 status_manager.go:851] "Failed to get status for pod" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4hqr\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.244027 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.244166 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.244296 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.244436 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247210 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247244 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247258 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247270 4809 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247283 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6p4v\" (UniqueName: \"kubernetes.io/projected/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-kube-api-access-z6p4v\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247293 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247305 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247317 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247329 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247340 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247352 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.247367 4809 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.250794 4809 scope.go:117] "RemoveContainer" containerID="8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678" Dec 05 11:13:00 crc kubenswrapper[4809]: E1205 11:13:00.251408 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678\": container with ID starting with 8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678 not found: ID does not exist" containerID="8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678" Dec 05 11:13:00 crc kubenswrapper[4809]: I1205 11:13:00.251435 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678"} err="failed to get container status \"8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678\": rpc error: code = NotFound desc = could not find container \"8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678\": container with ID starting with 8c3765d93f85b3ff5d2ab375be34a7f637ed5e6b47e895293a9f8a7a606df678 not found: ID does not exist" Dec 05 11:13:00 crc kubenswrapper[4809]: E1205 11:13:00.643449 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Dec 05 11:13:02 crc kubenswrapper[4809]: E1205 11:13:02.244235 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="3.2s" Dec 05 11:13:04 crc kubenswrapper[4809]: I1205 11:13:04.875047 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:04.875824 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:04.876282 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:04.876524 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:04.876791 4809 status_manager.go:851] "Failed to get status for pod" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4hqr\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: E1205 11:13:05.445586 4809 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="6.4s" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.872084 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.873114 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.873906 4809 status_manager.go:851] "Failed to get status for pod" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4hqr\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.874505 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.875028 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.875459 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.890125 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.890160 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:05 crc kubenswrapper[4809]: E1205 11:13:05.890700 4809 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:05 crc kubenswrapper[4809]: I1205 11:13:05.891310 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:06 crc kubenswrapper[4809]: I1205 11:13:06.258493 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"300258da7327aa77cf0fc1421c5012f09bb696f4824e65c05f656a1c4cdbcba5"} Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.268697 4809 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="095a78f38ae316fecd73cb783b0ea937d951c030048ce612f86fabcd5e2e3307" exitCode=0 Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.268815 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"095a78f38ae316fecd73cb783b0ea937d951c030048ce612f86fabcd5e2e3307"} Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.269107 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.269126 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.269456 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: E1205 11:13:07.269760 4809 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.270007 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.270400 4809 status_manager.go:851] "Failed to get status for pod" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4hqr\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.270803 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.271347 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.272183 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.272242 4809 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b" exitCode=1 Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.272274 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b"} Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.272660 4809 scope.go:117] "RemoveContainer" containerID="1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.272914 4809 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.273455 4809 status_manager.go:851] "Failed to get status for pod" podUID="b6f405ac-b705-477f-b150-109df1092b2b" pod="openshift-marketplace/redhat-operators-hlxqf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hlxqf\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.273758 4809 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.273996 4809 status_manager.go:851] "Failed to get status for pod" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" pod="openshift-authentication/oauth-openshift-558db77b4-t4hqr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-t4hqr\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.274190 4809 status_manager.go:851] "Failed to get status for pod" podUID="bbdca384-2491-4744-913f-8be714d865a1" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:07 crc kubenswrapper[4809]: I1205 11:13:07.274404 4809 status_manager.go:851] "Failed to get status for pod" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" pod="openshift-marketplace/redhat-marketplace-dmzbq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dmzbq\": dial tcp 38.102.83.180:6443: connect: connection refused" Dec 05 11:13:08 crc kubenswrapper[4809]: I1205 11:13:08.299727 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"94ea1aefc746e5e37fa3cca756cd96955b09722972f04a182b1d99890fcbd85b"} Dec 05 11:13:08 crc kubenswrapper[4809]: I1205 11:13:08.300556 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f8fcab9048e911562df44234332f4a6764fb3209b01702c016e91fbb75b31bd2"} Dec 05 11:13:08 crc kubenswrapper[4809]: I1205 11:13:08.300574 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a9c1be1a500a7264c1f69be654be0a53328bb421c17bede9c1661c39d51bee76"} Dec 05 11:13:08 crc kubenswrapper[4809]: I1205 11:13:08.312138 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 11:13:08 crc kubenswrapper[4809]: I1205 11:13:08.312215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"28cf389d273fa797aeb44e5780d2ca01f0689d1f7b6169c7fdaba586640fa626"} Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.019647 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.321539 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6e157903ddb2c6fd994536264147fd243874c88bf946d8cd8e4bd4b5a3c5f6b5"} Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.321600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"aac6f5250462063ea57d9b6ad03b5956e5d84410b45a9bb18f80b527c80fb2f7"} Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.322007 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.322026 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.991782 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.992072 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 11:13:09 crc kubenswrapper[4809]: I1205 11:13:09.992125 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 11:13:10 crc kubenswrapper[4809]: I1205 11:13:10.893130 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:10 crc kubenswrapper[4809]: I1205 11:13:10.894782 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:10 crc kubenswrapper[4809]: I1205 11:13:10.900986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:14 crc kubenswrapper[4809]: I1205 11:13:14.412369 4809 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:14 crc kubenswrapper[4809]: I1205 11:13:14.900739 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="edaaf61e-cb7e-4b7e-b820-9e9d0b616bc9" Dec 05 11:13:15 crc kubenswrapper[4809]: I1205 11:13:15.351088 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:15 crc kubenswrapper[4809]: I1205 11:13:15.351190 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:15 crc kubenswrapper[4809]: I1205 11:13:15.351222 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:15 crc kubenswrapper[4809]: I1205 11:13:15.357578 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="edaaf61e-cb7e-4b7e-b820-9e9d0b616bc9" Dec 05 11:13:16 crc kubenswrapper[4809]: I1205 11:13:16.357896 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:16 crc kubenswrapper[4809]: I1205 11:13:16.357940 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:16 crc kubenswrapper[4809]: I1205 11:13:16.362262 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="edaaf61e-cb7e-4b7e-b820-9e9d0b616bc9" Dec 05 11:13:16 crc kubenswrapper[4809]: I1205 11:13:16.364086 4809 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://a9c1be1a500a7264c1f69be654be0a53328bb421c17bede9c1661c39d51bee76" Dec 05 11:13:16 crc kubenswrapper[4809]: I1205 11:13:16.364116 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:17 crc kubenswrapper[4809]: I1205 11:13:17.363319 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:17 crc kubenswrapper[4809]: I1205 11:13:17.363361 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:17 crc kubenswrapper[4809]: I1205 11:13:17.368209 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="edaaf61e-cb7e-4b7e-b820-9e9d0b616bc9" Dec 05 11:13:19 crc kubenswrapper[4809]: I1205 11:13:19.991919 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 11:13:19 crc kubenswrapper[4809]: I1205 11:13:19.992238 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 11:13:24 crc kubenswrapper[4809]: I1205 11:13:24.017126 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 11:13:24 crc kubenswrapper[4809]: I1205 11:13:24.646971 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 11:13:24 crc kubenswrapper[4809]: I1205 11:13:24.944815 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 11:13:25 crc kubenswrapper[4809]: I1205 11:13:25.281439 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 11:13:25 crc kubenswrapper[4809]: I1205 11:13:25.330269 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 11:13:25 crc kubenswrapper[4809]: I1205 11:13:25.401781 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 11:13:25 crc kubenswrapper[4809]: I1205 11:13:25.621358 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 11:13:25 crc kubenswrapper[4809]: I1205 11:13:25.799967 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 11:13:25 crc kubenswrapper[4809]: I1205 11:13:25.859416 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.009875 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.131747 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.221208 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.537675 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.565495 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.721684 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.727624 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.809785 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 11:13:26 crc kubenswrapper[4809]: I1205 11:13:26.846766 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.140930 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.144152 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.150147 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.156008 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.175908 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.233572 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.284490 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.393764 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.412896 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.612537 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.646293 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.667262 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.683620 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.694884 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.851677 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.877237 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.889355 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 11:13:27 crc kubenswrapper[4809]: I1205 11:13:27.923089 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.100597 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.113844 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.136408 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.230746 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.304599 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.411921 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.488545 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.496445 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.602295 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.618065 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.697496 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.704658 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.738168 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 11:13:28 crc kubenswrapper[4809]: I1205 11:13:28.785325 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.011949 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.085961 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.240538 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.247739 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.279254 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.356440 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.424799 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.498980 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.500561 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.575286 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.590013 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.717602 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.847815 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.854932 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.916192 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.939726 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.967694 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.991601 4809 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.991688 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.991749 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.991908 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.992577 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"28cf389d273fa797aeb44e5780d2ca01f0689d1f7b6169c7fdaba586640fa626"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 05 11:13:29 crc kubenswrapper[4809]: I1205 11:13:29.992765 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://28cf389d273fa797aeb44e5780d2ca01f0689d1f7b6169c7fdaba586640fa626" gracePeriod=30 Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.008942 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.017037 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.031220 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.098539 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.143535 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.199689 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.393857 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.395923 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.418529 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.467617 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.484760 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.496904 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.523755 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.578651 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.681826 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.757066 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.796390 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.820476 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.841229 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 11:13:30 crc kubenswrapper[4809]: I1205 11:13:30.958162 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.032425 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.064190 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.155125 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.197591 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.285801 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.316781 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.325095 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.343444 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.395513 4809 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.413605 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.461688 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.520440 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.537878 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.809564 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.835352 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 11:13:31 crc kubenswrapper[4809]: I1205 11:13:31.893801 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.056956 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.060409 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.103623 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.127339 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.204804 4809 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.210999 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.210976726 podStartE2EDuration="40.210976726s" podCreationTimestamp="2025-12-05 11:12:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:13:14.47430027 +0000 UTC m=+289.865276838" watchObservedRunningTime="2025-12-05 11:13:32.210976726 +0000 UTC m=+307.601953274" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.212625 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmzbq","openshift-marketplace/redhat-operators-hlxqf","openshift-authentication/oauth-openshift-558db77b4-t4hqr","openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.212724 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-68974c876c-rrshm","openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213002 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="registry-server" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213028 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="registry-server" Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213044 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="extract-utilities" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213053 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="extract-utilities" Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213070 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" containerName="oauth-openshift" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213077 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" containerName="oauth-openshift" Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213087 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="registry-server" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213094 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="registry-server" Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213106 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="extract-utilities" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213115 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="extract-utilities" Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213126 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="extract-content" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213133 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="extract-content" Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213147 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbdca384-2491-4744-913f-8be714d865a1" containerName="installer" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213155 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbdca384-2491-4744-913f-8be714d865a1" containerName="installer" Dec 05 11:13:32 crc kubenswrapper[4809]: E1205 11:13:32.213168 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="extract-content" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213175 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="extract-content" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213222 4809 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213256 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2e8da43f-5236-4877-9010-364b45224889" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213326 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" containerName="registry-server" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213345 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" containerName="oauth-openshift" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213356 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f405ac-b705-477f-b150-109df1092b2b" containerName="registry-server" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213372 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbdca384-2491-4744-913f-8be714d865a1" containerName="installer" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.213916 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.216925 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.217878 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.217912 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.219685 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.219840 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.219839 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.220419 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.220521 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.222273 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.224285 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.224418 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.224586 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.224728 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.228047 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.234170 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.239299 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.251231 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.251209957 podStartE2EDuration="18.251209957s" podCreationTimestamp="2025-12-05 11:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:13:32.247480691 +0000 UTC m=+307.638457269" watchObservedRunningTime="2025-12-05 11:13:32.251209957 +0000 UTC m=+307.642186515" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.280168 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.324621 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.346925 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.370570 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.370662 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.370707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-router-certs\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.370729 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.370759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-session\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.370783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.370814 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.371137 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-error\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.371182 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-audit-policies\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.371218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-audit-dir\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.371246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.371284 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-service-ca\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.371313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-login\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.371358 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhqtv\" (UniqueName: \"kubernetes.io/projected/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-kube-api-access-jhqtv\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.412977 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.414720 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.444458 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.469493 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-service-ca\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472296 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-login\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472344 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhqtv\" (UniqueName: \"kubernetes.io/projected/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-kube-api-access-jhqtv\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472450 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-router-certs\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472474 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-session\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.472742 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.473439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-error\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.473485 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-audit-policies\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.473530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-audit-dir\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.473573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-service-ca\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.473603 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-audit-dir\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.473781 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.474194 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.474230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-audit-policies\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.479716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-login\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.479727 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.479728 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.480379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.481803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-router-certs\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.481889 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-template-error\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.483286 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.483592 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-v4-0-config-system-session\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.488788 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhqtv\" (UniqueName: \"kubernetes.io/projected/ff5b7ffc-82da-4ab4-a0be-45289d3c6870-kube-api-access-jhqtv\") pod \"oauth-openshift-68974c876c-rrshm\" (UID: \"ff5b7ffc-82da-4ab4-a0be-45289d3c6870\") " pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.517075 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.537157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.588772 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.602906 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.620093 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.627829 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.645274 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.655114 4809 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.655177 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.785388 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.793771 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.881432 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ac8b761-cb00-4b83-bc1e-0da82ee5c88b" path="/var/lib/kubelet/pods/1ac8b761-cb00-4b83-bc1e-0da82ee5c88b/volumes" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.882339 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73486c5f-7eeb-4a30-ad5c-2eaf64837d28" path="/var/lib/kubelet/pods/73486c5f-7eeb-4a30-ad5c-2eaf64837d28/volumes" Dec 05 11:13:32 crc kubenswrapper[4809]: I1205 11:13:32.883249 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f405ac-b705-477f-b150-109df1092b2b" path="/var/lib/kubelet/pods/b6f405ac-b705-477f-b150-109df1092b2b/volumes" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.053860 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.229208 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.230228 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.235001 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.262501 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.292724 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.322856 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.336817 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.433308 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.436696 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.472371 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.511723 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.658011 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.673070 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.817831 4809 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.889685 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.894192 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.900799 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.935849 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.965676 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.971836 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 11:13:33 crc kubenswrapper[4809]: I1205 11:13:33.985694 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.027260 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.106128 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.131326 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.159920 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.219204 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.233416 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.263110 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.264041 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.277119 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.343311 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.370156 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.446856 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.506942 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.532539 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.539296 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.620356 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.662674 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.673280 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.775742 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.883932 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.954417 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 11:13:34 crc kubenswrapper[4809]: I1205 11:13:34.986169 4809 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.053425 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.113981 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.164521 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.203815 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68974c876c-rrshm"] Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.249727 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.249736 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.296479 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.311422 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.345476 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.388759 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68974c876c-rrshm"] Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.465785 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" event={"ID":"ff5b7ffc-82da-4ab4-a0be-45289d3c6870","Type":"ContainerStarted","Data":"3a034f098537e8df974d3b7e771bd98513141268a7adcad6081e6def709ccd96"} Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.499567 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.646074 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.757938 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.792605 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.830220 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.864351 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.871111 4809 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.871340 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe" gracePeriod=5 Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.934712 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 11:13:35 crc kubenswrapper[4809]: I1205 11:13:35.961769 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.050973 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.057285 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.070173 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.242346 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.313459 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.426332 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.462342 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.472247 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" event={"ID":"ff5b7ffc-82da-4ab4-a0be-45289d3c6870","Type":"ContainerStarted","Data":"2d72b7469db2726d7b970ed85714fbd7986f105dffa5826b4769d02718717249"} Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.472497 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.478016 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.496869 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-68974c876c-rrshm" podStartSLOduration=62.496849081 podStartE2EDuration="1m2.496849081s" podCreationTimestamp="2025-12-05 11:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:13:36.493075354 +0000 UTC m=+311.884051952" watchObservedRunningTime="2025-12-05 11:13:36.496849081 +0000 UTC m=+311.887825639" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.548920 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.751981 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.864361 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 11:13:36 crc kubenswrapper[4809]: I1205 11:13:36.935256 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.056982 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.413004 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.454976 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.498891 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.542979 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.656674 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.685260 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.692604 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.720933 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.740298 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.846612 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 11:13:37 crc kubenswrapper[4809]: I1205 11:13:37.974433 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.019787 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.140773 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.321362 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.499473 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.564869 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.638546 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.639211 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 11:13:38 crc kubenswrapper[4809]: I1205 11:13:38.721565 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.163911 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.185509 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.213752 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.236193 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.241336 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.422592 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.445480 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.640793 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.714479 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.891041 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 11:13:39 crc kubenswrapper[4809]: I1205 11:13:39.998107 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 11:13:40 crc kubenswrapper[4809]: I1205 11:13:40.167385 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 11:13:40 crc kubenswrapper[4809]: I1205 11:13:40.428456 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.072173 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.446162 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.446234 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.502449 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.502514 4809 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe" exitCode=137 Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.502569 4809 scope.go:117] "RemoveContainer" containerID="776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.502599 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.517853 4809 scope.go:117] "RemoveContainer" containerID="776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe" Dec 05 11:13:41 crc kubenswrapper[4809]: E1205 11:13:41.518234 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe\": container with ID starting with 776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe not found: ID does not exist" containerID="776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.518277 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe"} err="failed to get container status \"776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe\": rpc error: code = NotFound desc = could not find container \"776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe\": container with ID starting with 776479da64b9d1e133316a43db890a0739b06d103781ff19c5866b766cb211fe not found: ID does not exist" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.582993 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583067 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583131 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583166 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583210 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583885 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583936 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.583979 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.592828 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.684517 4809 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.684546 4809 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.684555 4809 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.684564 4809 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:41 crc kubenswrapper[4809]: I1205 11:13:41.684573 4809 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 11:13:42 crc kubenswrapper[4809]: I1205 11:13:42.881802 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 05 11:13:42 crc kubenswrapper[4809]: I1205 11:13:42.882484 4809 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 05 11:13:42 crc kubenswrapper[4809]: I1205 11:13:42.894240 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 11:13:42 crc kubenswrapper[4809]: I1205 11:13:42.894331 4809 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="1bed1408-c3c2-4b6b-9f47-3e4a88e02eb1" Dec 05 11:13:42 crc kubenswrapper[4809]: I1205 11:13:42.899082 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 11:13:42 crc kubenswrapper[4809]: I1205 11:13:42.899127 4809 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="1bed1408-c3c2-4b6b-9f47-3e4a88e02eb1" Dec 05 11:13:50 crc kubenswrapper[4809]: I1205 11:13:50.615556 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 11:13:59 crc kubenswrapper[4809]: I1205 11:13:59.438554 4809 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 11:14:00 crc kubenswrapper[4809]: I1205 11:14:00.609461 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 05 11:14:00 crc kubenswrapper[4809]: I1205 11:14:00.612098 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 11:14:00 crc kubenswrapper[4809]: I1205 11:14:00.612164 4809 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="28cf389d273fa797aeb44e5780d2ca01f0689d1f7b6169c7fdaba586640fa626" exitCode=137 Dec 05 11:14:00 crc kubenswrapper[4809]: I1205 11:14:00.612206 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"28cf389d273fa797aeb44e5780d2ca01f0689d1f7b6169c7fdaba586640fa626"} Dec 05 11:14:00 crc kubenswrapper[4809]: I1205 11:14:00.612261 4809 scope.go:117] "RemoveContainer" containerID="1e1874f14b9c9de9fe7277b069c4c77ee49b48c07c3b02996d8e4768240ff25b" Dec 05 11:14:01 crc kubenswrapper[4809]: I1205 11:14:01.618073 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 05 11:14:01 crc kubenswrapper[4809]: I1205 11:14:01.618935 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cb595c6788ec42a8098816cf5b9aa3be3adee0cab485810aab50dd82bc1cf4a6"} Dec 05 11:14:02 crc kubenswrapper[4809]: I1205 11:14:02.627083 4809 generic.go:334] "Generic (PLEG): container finished" podID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerID="1a746b32726ed8c572c8ecf84ccee922f17e05b980e33563e411bc7a0f460d7e" exitCode=0 Dec 05 11:14:02 crc kubenswrapper[4809]: I1205 11:14:02.627187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" event={"ID":"8781fe56-ed07-4998-b905-7fd3fc750e42","Type":"ContainerDied","Data":"1a746b32726ed8c572c8ecf84ccee922f17e05b980e33563e411bc7a0f460d7e"} Dec 05 11:14:02 crc kubenswrapper[4809]: I1205 11:14:02.627965 4809 scope.go:117] "RemoveContainer" containerID="1a746b32726ed8c572c8ecf84ccee922f17e05b980e33563e411bc7a0f460d7e" Dec 05 11:14:03 crc kubenswrapper[4809]: I1205 11:14:03.639303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" event={"ID":"8781fe56-ed07-4998-b905-7fd3fc750e42","Type":"ContainerStarted","Data":"0c70a1f8ffbd26b6c7010cb1914a3f5ebbddb2881f44cd4598b78da3d444bc60"} Dec 05 11:14:03 crc kubenswrapper[4809]: I1205 11:14:03.641386 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:14:03 crc kubenswrapper[4809]: I1205 11:14:03.643926 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:14:04 crc kubenswrapper[4809]: I1205 11:14:04.722827 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 11:14:06 crc kubenswrapper[4809]: I1205 11:14:06.219821 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 11:14:09 crc kubenswrapper[4809]: I1205 11:14:09.019707 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:14:09 crc kubenswrapper[4809]: I1205 11:14:09.246209 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 11:14:09 crc kubenswrapper[4809]: I1205 11:14:09.991372 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:14:09 crc kubenswrapper[4809]: I1205 11:14:09.997929 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:14:10 crc kubenswrapper[4809]: I1205 11:14:10.678131 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 11:14:10 crc kubenswrapper[4809]: I1205 11:14:10.938143 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.414367 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76"] Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.415245 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" podUID="677864f6-90af-4ffa-9b36-729194cedcd5" containerName="route-controller-manager" containerID="cri-o://899a2699a3be2d750c5149359415a2fae24c1871ea328ddab5cdc861d437e58a" gracePeriod=30 Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.418868 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gvwf"] Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.419105 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" podUID="b52152a3-7674-489f-b67b-e96346e88228" containerName="controller-manager" containerID="cri-o://65e97a7c38f9955063e988e7b6b2bb3f3c63535a0a25c0b54b0b304393df7495" gracePeriod=30 Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.718361 4809 generic.go:334] "Generic (PLEG): container finished" podID="677864f6-90af-4ffa-9b36-729194cedcd5" containerID="899a2699a3be2d750c5149359415a2fae24c1871ea328ddab5cdc861d437e58a" exitCode=0 Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.718414 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" event={"ID":"677864f6-90af-4ffa-9b36-729194cedcd5","Type":"ContainerDied","Data":"899a2699a3be2d750c5149359415a2fae24c1871ea328ddab5cdc861d437e58a"} Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.719818 4809 generic.go:334] "Generic (PLEG): container finished" podID="b52152a3-7674-489f-b67b-e96346e88228" containerID="65e97a7c38f9955063e988e7b6b2bb3f3c63535a0a25c0b54b0b304393df7495" exitCode=0 Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.719844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" event={"ID":"b52152a3-7674-489f-b67b-e96346e88228","Type":"ContainerDied","Data":"65e97a7c38f9955063e988e7b6b2bb3f3c63535a0a25c0b54b0b304393df7495"} Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.795504 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.869391 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.965757 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677864f6-90af-4ffa-9b36-729194cedcd5-serving-cert\") pod \"677864f6-90af-4ffa-9b36-729194cedcd5\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.965906 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-client-ca\") pod \"677864f6-90af-4ffa-9b36-729194cedcd5\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.966457 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-client-ca" (OuterVolumeSpecName: "client-ca") pod "677864f6-90af-4ffa-9b36-729194cedcd5" (UID: "677864f6-90af-4ffa-9b36-729194cedcd5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.966533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-config\") pod \"677864f6-90af-4ffa-9b36-729194cedcd5\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.967046 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-config" (OuterVolumeSpecName: "config") pod "677864f6-90af-4ffa-9b36-729194cedcd5" (UID: "677864f6-90af-4ffa-9b36-729194cedcd5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.967452 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl4z9\" (UniqueName: \"kubernetes.io/projected/677864f6-90af-4ffa-9b36-729194cedcd5-kube-api-access-jl4z9\") pod \"677864f6-90af-4ffa-9b36-729194cedcd5\" (UID: \"677864f6-90af-4ffa-9b36-729194cedcd5\") " Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.967524 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-config\") pod \"b52152a3-7674-489f-b67b-e96346e88228\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.967690 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.967708 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677864f6-90af-4ffa-9b36-729194cedcd5-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.968510 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-config" (OuterVolumeSpecName: "config") pod "b52152a3-7674-489f-b67b-e96346e88228" (UID: "b52152a3-7674-489f-b67b-e96346e88228"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.970859 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/677864f6-90af-4ffa-9b36-729194cedcd5-kube-api-access-jl4z9" (OuterVolumeSpecName: "kube-api-access-jl4z9") pod "677864f6-90af-4ffa-9b36-729194cedcd5" (UID: "677864f6-90af-4ffa-9b36-729194cedcd5"). InnerVolumeSpecName "kube-api-access-jl4z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:14:18 crc kubenswrapper[4809]: I1205 11:14:18.971037 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/677864f6-90af-4ffa-9b36-729194cedcd5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "677864f6-90af-4ffa-9b36-729194cedcd5" (UID: "677864f6-90af-4ffa-9b36-729194cedcd5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.068701 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-client-ca\") pod \"b52152a3-7674-489f-b67b-e96346e88228\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.068836 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8st7p\" (UniqueName: \"kubernetes.io/projected/b52152a3-7674-489f-b67b-e96346e88228-kube-api-access-8st7p\") pod \"b52152a3-7674-489f-b67b-e96346e88228\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.068892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b52152a3-7674-489f-b67b-e96346e88228-serving-cert\") pod \"b52152a3-7674-489f-b67b-e96346e88228\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.068953 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-proxy-ca-bundles\") pod \"b52152a3-7674-489f-b67b-e96346e88228\" (UID: \"b52152a3-7674-489f-b67b-e96346e88228\") " Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.069335 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl4z9\" (UniqueName: \"kubernetes.io/projected/677864f6-90af-4ffa-9b36-729194cedcd5-kube-api-access-jl4z9\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.069353 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.069363 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677864f6-90af-4ffa-9b36-729194cedcd5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.069382 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-client-ca" (OuterVolumeSpecName: "client-ca") pod "b52152a3-7674-489f-b67b-e96346e88228" (UID: "b52152a3-7674-489f-b67b-e96346e88228"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.069711 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b52152a3-7674-489f-b67b-e96346e88228" (UID: "b52152a3-7674-489f-b67b-e96346e88228"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.073006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b52152a3-7674-489f-b67b-e96346e88228-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b52152a3-7674-489f-b67b-e96346e88228" (UID: "b52152a3-7674-489f-b67b-e96346e88228"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.073895 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b52152a3-7674-489f-b67b-e96346e88228-kube-api-access-8st7p" (OuterVolumeSpecName: "kube-api-access-8st7p") pod "b52152a3-7674-489f-b67b-e96346e88228" (UID: "b52152a3-7674-489f-b67b-e96346e88228"). InnerVolumeSpecName "kube-api-access-8st7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.170430 4809 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.170479 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b52152a3-7674-489f-b67b-e96346e88228-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.170492 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8st7p\" (UniqueName: \"kubernetes.io/projected/b52152a3-7674-489f-b67b-e96346e88228-kube-api-access-8st7p\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.170538 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b52152a3-7674-489f-b67b-e96346e88228-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.505590 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-647ccd6888-dfghn"] Dec 05 11:14:19 crc kubenswrapper[4809]: E1205 11:14:19.506078 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677864f6-90af-4ffa-9b36-729194cedcd5" containerName="route-controller-manager" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.506103 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="677864f6-90af-4ffa-9b36-729194cedcd5" containerName="route-controller-manager" Dec 05 11:14:19 crc kubenswrapper[4809]: E1205 11:14:19.506128 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52152a3-7674-489f-b67b-e96346e88228" containerName="controller-manager" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.506140 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52152a3-7674-489f-b67b-e96346e88228" containerName="controller-manager" Dec 05 11:14:19 crc kubenswrapper[4809]: E1205 11:14:19.506162 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.506172 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.506309 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b52152a3-7674-489f-b67b-e96346e88228" containerName="controller-manager" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.506325 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="677864f6-90af-4ffa-9b36-729194cedcd5" containerName="route-controller-manager" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.506339 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.507021 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.509932 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54"] Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.510871 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.520760 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-647ccd6888-dfghn"] Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.523529 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54"] Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.676503 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4rrp\" (UniqueName: \"kubernetes.io/projected/def43963-f5aa-4768-b5dd-d3fa86b88d31-kube-api-access-t4rrp\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.676854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-config\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.676893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-client-ca\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.676940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def43963-f5aa-4768-b5dd-d3fa86b88d31-serving-cert\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.676966 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-config\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.677095 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-proxy-ca-bundles\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.677206 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-client-ca\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.677245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfwfx\" (UniqueName: \"kubernetes.io/projected/229e2369-822d-406a-9f6b-d55425378c34-kube-api-access-dfwfx\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.677368 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/229e2369-822d-406a-9f6b-d55425378c34-serving-cert\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.739973 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" event={"ID":"b52152a3-7674-489f-b67b-e96346e88228","Type":"ContainerDied","Data":"97e8c535f2af5be6701590b3b314ec1611aafa61d82314e650bfb641c03e751c"} Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.740015 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gvwf" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.740023 4809 scope.go:117] "RemoveContainer" containerID="65e97a7c38f9955063e988e7b6b2bb3f3c63535a0a25c0b54b0b304393df7495" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.741822 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" event={"ID":"677864f6-90af-4ffa-9b36-729194cedcd5","Type":"ContainerDied","Data":"0ab14695c81fc30b85012971dd3086c04ea7bb42296805c1ed19d66a496ed413"} Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.741904 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.758400 4809 scope.go:117] "RemoveContainer" containerID="899a2699a3be2d750c5149359415a2fae24c1871ea328ddab5cdc861d437e58a" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.772482 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gvwf"] Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.777375 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gvwf"] Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.777955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-proxy-ca-bundles\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.777995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-client-ca\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.778016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfwfx\" (UniqueName: \"kubernetes.io/projected/229e2369-822d-406a-9f6b-d55425378c34-kube-api-access-dfwfx\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.778059 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/229e2369-822d-406a-9f6b-d55425378c34-serving-cert\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.778084 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4rrp\" (UniqueName: \"kubernetes.io/projected/def43963-f5aa-4768-b5dd-d3fa86b88d31-kube-api-access-t4rrp\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.778106 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-config\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.778123 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-client-ca\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.778138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def43963-f5aa-4768-b5dd-d3fa86b88d31-serving-cert\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.778152 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-config\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.779308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-client-ca\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.779746 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-proxy-ca-bundles\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.779943 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-config\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.781002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229e2369-822d-406a-9f6b-d55425378c34-config\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.784423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-client-ca\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.786493 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/229e2369-822d-406a-9f6b-d55425378c34-serving-cert\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.790810 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76"] Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.790861 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wlw76"] Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.793255 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def43963-f5aa-4768-b5dd-d3fa86b88d31-serving-cert\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.795202 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4rrp\" (UniqueName: \"kubernetes.io/projected/def43963-f5aa-4768-b5dd-d3fa86b88d31-kube-api-access-t4rrp\") pod \"route-controller-manager-84768cf8b-m6k54\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.796736 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfwfx\" (UniqueName: \"kubernetes.io/projected/229e2369-822d-406a-9f6b-d55425378c34-kube-api-access-dfwfx\") pod \"controller-manager-647ccd6888-dfghn\" (UID: \"229e2369-822d-406a-9f6b-d55425378c34\") " pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.866255 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:19 crc kubenswrapper[4809]: I1205 11:14:19.887723 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.135349 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54"] Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.272614 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-647ccd6888-dfghn"] Dec 05 11:14:20 crc kubenswrapper[4809]: W1205 11:14:20.277341 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod229e2369_822d_406a_9f6b_d55425378c34.slice/crio-497fbe7cacb17b84f8685cf393001e03bc9282d89e4478ae03734b0230eaf6b7 WatchSource:0}: Error finding container 497fbe7cacb17b84f8685cf393001e03bc9282d89e4478ae03734b0230eaf6b7: Status 404 returned error can't find the container with id 497fbe7cacb17b84f8685cf393001e03bc9282d89e4478ae03734b0230eaf6b7 Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.749111 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" event={"ID":"229e2369-822d-406a-9f6b-d55425378c34","Type":"ContainerStarted","Data":"8ee79f221643a1d5a6ef0eea5aeb157330537a31675e0a1553665da966446066"} Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.749457 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.749474 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" event={"ID":"229e2369-822d-406a-9f6b-d55425378c34","Type":"ContainerStarted","Data":"497fbe7cacb17b84f8685cf393001e03bc9282d89e4478ae03734b0230eaf6b7"} Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.750715 4809 patch_prober.go:28] interesting pod/controller-manager-647ccd6888-dfghn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.750769 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" podUID="229e2369-822d-406a-9f6b-d55425378c34" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.752314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" event={"ID":"def43963-f5aa-4768-b5dd-d3fa86b88d31","Type":"ContainerStarted","Data":"ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd"} Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.752359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" event={"ID":"def43963-f5aa-4768-b5dd-d3fa86b88d31","Type":"ContainerStarted","Data":"35f7f972d70ac15c09c562bbb7e8c17d7071f65743adc3ccb0f8b897ffd4de2e"} Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.752382 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.768998 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" podStartSLOduration=2.76897577 podStartE2EDuration="2.76897577s" podCreationTimestamp="2025-12-05 11:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:14:20.765774194 +0000 UTC m=+356.156750762" watchObservedRunningTime="2025-12-05 11:14:20.76897577 +0000 UTC m=+356.159952348" Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.781478 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" podStartSLOduration=2.781458945 podStartE2EDuration="2.781458945s" podCreationTimestamp="2025-12-05 11:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:14:20.780099435 +0000 UTC m=+356.171076003" watchObservedRunningTime="2025-12-05 11:14:20.781458945 +0000 UTC m=+356.172435503" Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.878603 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="677864f6-90af-4ffa-9b36-729194cedcd5" path="/var/lib/kubelet/pods/677864f6-90af-4ffa-9b36-729194cedcd5/volumes" Dec 05 11:14:20 crc kubenswrapper[4809]: I1205 11:14:20.879399 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b52152a3-7674-489f-b67b-e96346e88228" path="/var/lib/kubelet/pods/b52152a3-7674-489f-b67b-e96346e88228/volumes" Dec 05 11:14:21 crc kubenswrapper[4809]: I1205 11:14:21.193678 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:21 crc kubenswrapper[4809]: I1205 11:14:21.770425 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-647ccd6888-dfghn" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.064617 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54"] Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.065384 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" podUID="def43963-f5aa-4768-b5dd-d3fa86b88d31" containerName="route-controller-manager" containerID="cri-o://ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd" gracePeriod=30 Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.558904 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.752906 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-config\") pod \"def43963-f5aa-4768-b5dd-d3fa86b88d31\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.752998 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-client-ca\") pod \"def43963-f5aa-4768-b5dd-d3fa86b88d31\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.753125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4rrp\" (UniqueName: \"kubernetes.io/projected/def43963-f5aa-4768-b5dd-d3fa86b88d31-kube-api-access-t4rrp\") pod \"def43963-f5aa-4768-b5dd-d3fa86b88d31\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.753195 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def43963-f5aa-4768-b5dd-d3fa86b88d31-serving-cert\") pod \"def43963-f5aa-4768-b5dd-d3fa86b88d31\" (UID: \"def43963-f5aa-4768-b5dd-d3fa86b88d31\") " Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.753889 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-client-ca" (OuterVolumeSpecName: "client-ca") pod "def43963-f5aa-4768-b5dd-d3fa86b88d31" (UID: "def43963-f5aa-4768-b5dd-d3fa86b88d31"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.754014 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-config" (OuterVolumeSpecName: "config") pod "def43963-f5aa-4768-b5dd-d3fa86b88d31" (UID: "def43963-f5aa-4768-b5dd-d3fa86b88d31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.761488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/def43963-f5aa-4768-b5dd-d3fa86b88d31-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "def43963-f5aa-4768-b5dd-d3fa86b88d31" (UID: "def43963-f5aa-4768-b5dd-d3fa86b88d31"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.764105 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def43963-f5aa-4768-b5dd-d3fa86b88d31-kube-api-access-t4rrp" (OuterVolumeSpecName: "kube-api-access-t4rrp") pod "def43963-f5aa-4768-b5dd-d3fa86b88d31" (UID: "def43963-f5aa-4768-b5dd-d3fa86b88d31"). InnerVolumeSpecName "kube-api-access-t4rrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.854424 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4rrp\" (UniqueName: \"kubernetes.io/projected/def43963-f5aa-4768-b5dd-d3fa86b88d31-kube-api-access-t4rrp\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.854453 4809 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def43963-f5aa-4768-b5dd-d3fa86b88d31-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.854469 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.854487 4809 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/def43963-f5aa-4768-b5dd-d3fa86b88d31-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.856951 4809 generic.go:334] "Generic (PLEG): container finished" podID="def43963-f5aa-4768-b5dd-d3fa86b88d31" containerID="ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd" exitCode=0 Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.857004 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" event={"ID":"def43963-f5aa-4768-b5dd-d3fa86b88d31","Type":"ContainerDied","Data":"ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd"} Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.857037 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.857068 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54" event={"ID":"def43963-f5aa-4768-b5dd-d3fa86b88d31","Type":"ContainerDied","Data":"35f7f972d70ac15c09c562bbb7e8c17d7071f65743adc3ccb0f8b897ffd4de2e"} Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.857117 4809 scope.go:117] "RemoveContainer" containerID="ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.873924 4809 scope.go:117] "RemoveContainer" containerID="ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd" Dec 05 11:14:40 crc kubenswrapper[4809]: E1205 11:14:40.874323 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd\": container with ID starting with ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd not found: ID does not exist" containerID="ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.874352 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd"} err="failed to get container status \"ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd\": rpc error: code = NotFound desc = could not find container \"ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd\": container with ID starting with ecd4ea8713c5811bd1e8e32b39d0adc4d97afe9fd663d34cdc4335400d3479cd not found: ID does not exist" Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.890067 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54"] Dec 05 11:14:40 crc kubenswrapper[4809]: I1205 11:14:40.893362 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84768cf8b-m6k54"] Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.520777 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc"] Dec 05 11:14:41 crc kubenswrapper[4809]: E1205 11:14:41.520995 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def43963-f5aa-4768-b5dd-d3fa86b88d31" containerName="route-controller-manager" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.521033 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="def43963-f5aa-4768-b5dd-d3fa86b88d31" containerName="route-controller-manager" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.521197 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="def43963-f5aa-4768-b5dd-d3fa86b88d31" containerName="route-controller-manager" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.521668 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.524055 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.524362 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.524623 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.524723 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.524836 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.525290 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.528443 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc"] Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.664887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a577150-ebac-4d94-9972-5a206e5caebd-client-ca\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.664951 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8gsj\" (UniqueName: \"kubernetes.io/projected/4a577150-ebac-4d94-9972-5a206e5caebd-kube-api-access-f8gsj\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.664991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a577150-ebac-4d94-9972-5a206e5caebd-serving-cert\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.665054 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a577150-ebac-4d94-9972-5a206e5caebd-config\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.765915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a577150-ebac-4d94-9972-5a206e5caebd-config\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.766220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a577150-ebac-4d94-9972-5a206e5caebd-client-ca\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.766437 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8gsj\" (UniqueName: \"kubernetes.io/projected/4a577150-ebac-4d94-9972-5a206e5caebd-kube-api-access-f8gsj\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.766622 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a577150-ebac-4d94-9972-5a206e5caebd-serving-cert\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.767633 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a577150-ebac-4d94-9972-5a206e5caebd-config\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.767756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a577150-ebac-4d94-9972-5a206e5caebd-client-ca\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.779272 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a577150-ebac-4d94-9972-5a206e5caebd-serving-cert\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.787549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8gsj\" (UniqueName: \"kubernetes.io/projected/4a577150-ebac-4d94-9972-5a206e5caebd-kube-api-access-f8gsj\") pod \"route-controller-manager-55494645dd-p5nmc\" (UID: \"4a577150-ebac-4d94-9972-5a206e5caebd\") " pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:41 crc kubenswrapper[4809]: I1205 11:14:41.845444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:42 crc kubenswrapper[4809]: I1205 11:14:42.232915 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc"] Dec 05 11:14:42 crc kubenswrapper[4809]: W1205 11:14:42.240519 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a577150_ebac_4d94_9972_5a206e5caebd.slice/crio-9bc617440c78c2c4e997f50707b6a27bd867f30b9c0304ae1e4ab94a1bd7f463 WatchSource:0}: Error finding container 9bc617440c78c2c4e997f50707b6a27bd867f30b9c0304ae1e4ab94a1bd7f463: Status 404 returned error can't find the container with id 9bc617440c78c2c4e997f50707b6a27bd867f30b9c0304ae1e4ab94a1bd7f463 Dec 05 11:14:42 crc kubenswrapper[4809]: I1205 11:14:42.868042 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" event={"ID":"4a577150-ebac-4d94-9972-5a206e5caebd","Type":"ContainerStarted","Data":"2bd0b57f456095e0eefc423e1b6c2bb7a885769600974f4d25787aad34d58846"} Dec 05 11:14:42 crc kubenswrapper[4809]: I1205 11:14:42.868386 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:42 crc kubenswrapper[4809]: I1205 11:14:42.868403 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" event={"ID":"4a577150-ebac-4d94-9972-5a206e5caebd","Type":"ContainerStarted","Data":"9bc617440c78c2c4e997f50707b6a27bd867f30b9c0304ae1e4ab94a1bd7f463"} Dec 05 11:14:42 crc kubenswrapper[4809]: I1205 11:14:42.878741 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="def43963-f5aa-4768-b5dd-d3fa86b88d31" path="/var/lib/kubelet/pods/def43963-f5aa-4768-b5dd-d3fa86b88d31/volumes" Dec 05 11:14:42 crc kubenswrapper[4809]: I1205 11:14:42.879461 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" Dec 05 11:14:42 crc kubenswrapper[4809]: I1205 11:14:42.894005 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55494645dd-p5nmc" podStartSLOduration=2.8939813880000003 podStartE2EDuration="2.893981388s" podCreationTimestamp="2025-12-05 11:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:14:42.888212944 +0000 UTC m=+378.279189522" watchObservedRunningTime="2025-12-05 11:14:42.893981388 +0000 UTC m=+378.284957966" Dec 05 11:14:44 crc kubenswrapper[4809]: I1205 11:14:44.046379 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:14:44 crc kubenswrapper[4809]: I1205 11:14:44.046708 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.151927 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg"] Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.153274 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.155163 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.158325 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.165474 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg"] Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.249958 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56f75595-2600-4f31-9204-c94f4d6e212f-config-volume\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.250031 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56f75595-2600-4f31-9204-c94f4d6e212f-secret-volume\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.250069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmf8x\" (UniqueName: \"kubernetes.io/projected/56f75595-2600-4f31-9204-c94f4d6e212f-kube-api-access-fmf8x\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.350824 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56f75595-2600-4f31-9204-c94f4d6e212f-config-volume\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.350894 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56f75595-2600-4f31-9204-c94f4d6e212f-secret-volume\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.350927 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmf8x\" (UniqueName: \"kubernetes.io/projected/56f75595-2600-4f31-9204-c94f4d6e212f-kube-api-access-fmf8x\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.352051 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56f75595-2600-4f31-9204-c94f4d6e212f-config-volume\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.359264 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56f75595-2600-4f31-9204-c94f4d6e212f-secret-volume\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.371905 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmf8x\" (UniqueName: \"kubernetes.io/projected/56f75595-2600-4f31-9204-c94f4d6e212f-kube-api-access-fmf8x\") pod \"collect-profiles-29415555-jcsgg\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.483216 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.678362 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg"] Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.966222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" event={"ID":"56f75595-2600-4f31-9204-c94f4d6e212f","Type":"ContainerStarted","Data":"b700cfa536a2b49e4d9c37807eab2f08d31a79547d9f084f54e09ec0ed45057f"} Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.966265 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" event={"ID":"56f75595-2600-4f31-9204-c94f4d6e212f","Type":"ContainerStarted","Data":"a495cba76994b9cbe49bd72ef6e03bb8282096dc00441d350e77e6af43524fc8"} Dec 05 11:15:00 crc kubenswrapper[4809]: I1205 11:15:00.986782 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" podStartSLOduration=0.986764526 podStartE2EDuration="986.764526ms" podCreationTimestamp="2025-12-05 11:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:15:00.983299479 +0000 UTC m=+396.374276057" watchObservedRunningTime="2025-12-05 11:15:00.986764526 +0000 UTC m=+396.377741084" Dec 05 11:15:01 crc kubenswrapper[4809]: I1205 11:15:01.975619 4809 generic.go:334] "Generic (PLEG): container finished" podID="56f75595-2600-4f31-9204-c94f4d6e212f" containerID="b700cfa536a2b49e4d9c37807eab2f08d31a79547d9f084f54e09ec0ed45057f" exitCode=0 Dec 05 11:15:01 crc kubenswrapper[4809]: I1205 11:15:01.975971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" event={"ID":"56f75595-2600-4f31-9204-c94f4d6e212f","Type":"ContainerDied","Data":"b700cfa536a2b49e4d9c37807eab2f08d31a79547d9f084f54e09ec0ed45057f"} Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.216269 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.388880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56f75595-2600-4f31-9204-c94f4d6e212f-config-volume\") pod \"56f75595-2600-4f31-9204-c94f4d6e212f\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.389204 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmf8x\" (UniqueName: \"kubernetes.io/projected/56f75595-2600-4f31-9204-c94f4d6e212f-kube-api-access-fmf8x\") pod \"56f75595-2600-4f31-9204-c94f4d6e212f\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.389240 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56f75595-2600-4f31-9204-c94f4d6e212f-secret-volume\") pod \"56f75595-2600-4f31-9204-c94f4d6e212f\" (UID: \"56f75595-2600-4f31-9204-c94f4d6e212f\") " Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.389421 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56f75595-2600-4f31-9204-c94f4d6e212f-config-volume" (OuterVolumeSpecName: "config-volume") pod "56f75595-2600-4f31-9204-c94f4d6e212f" (UID: "56f75595-2600-4f31-9204-c94f4d6e212f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.389594 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56f75595-2600-4f31-9204-c94f4d6e212f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.396388 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f75595-2600-4f31-9204-c94f4d6e212f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "56f75595-2600-4f31-9204-c94f4d6e212f" (UID: "56f75595-2600-4f31-9204-c94f4d6e212f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.397084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56f75595-2600-4f31-9204-c94f4d6e212f-kube-api-access-fmf8x" (OuterVolumeSpecName: "kube-api-access-fmf8x") pod "56f75595-2600-4f31-9204-c94f4d6e212f" (UID: "56f75595-2600-4f31-9204-c94f4d6e212f"). InnerVolumeSpecName "kube-api-access-fmf8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.493032 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmf8x\" (UniqueName: \"kubernetes.io/projected/56f75595-2600-4f31-9204-c94f4d6e212f-kube-api-access-fmf8x\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.493080 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56f75595-2600-4f31-9204-c94f4d6e212f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.990497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" event={"ID":"56f75595-2600-4f31-9204-c94f4d6e212f","Type":"ContainerDied","Data":"a495cba76994b9cbe49bd72ef6e03bb8282096dc00441d350e77e6af43524fc8"} Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.990546 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a495cba76994b9cbe49bd72ef6e03bb8282096dc00441d350e77e6af43524fc8" Dec 05 11:15:03 crc kubenswrapper[4809]: I1205 11:15:03.990555 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg" Dec 05 11:15:14 crc kubenswrapper[4809]: I1205 11:15:14.046372 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:15:14 crc kubenswrapper[4809]: I1205 11:15:14.046962 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.801319 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9t5j4"] Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.803959 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9t5j4" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="registry-server" containerID="cri-o://cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74" gracePeriod=30 Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.808689 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xj244"] Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.808953 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xj244" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="registry-server" containerID="cri-o://74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de" gracePeriod=30 Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.818591 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wfbxx"] Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.818857 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" containerID="cri-o://0c70a1f8ffbd26b6c7010cb1914a3f5ebbddb2881f44cd4598b78da3d444bc60" gracePeriod=30 Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.828693 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcp4f"] Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.828962 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qcp4f" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="registry-server" containerID="cri-o://4a4a8a556e52e0d31f19203c359c066f8865ba75e234a413712498d815eb0b00" gracePeriod=30 Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.847750 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t4mbz"] Dec 05 11:15:23 crc kubenswrapper[4809]: E1205 11:15:23.848125 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f75595-2600-4f31-9204-c94f4d6e212f" containerName="collect-profiles" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.848152 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f75595-2600-4f31-9204-c94f4d6e212f" containerName="collect-profiles" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.848329 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f75595-2600-4f31-9204-c94f4d6e212f" containerName="collect-profiles" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.848922 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.866861 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jrg7v"] Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.867418 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jrg7v" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="registry-server" containerID="cri-o://1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92" gracePeriod=30 Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.876015 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t4mbz"] Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.895983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.896046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7ksn\" (UniqueName: \"kubernetes.io/projected/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-kube-api-access-n7ksn\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.896082 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.996889 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.996933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7ksn\" (UniqueName: \"kubernetes.io/projected/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-kube-api-access-n7ksn\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.996973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:23 crc kubenswrapper[4809]: I1205 11:15:23.997787 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.003213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.010718 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7ksn\" (UniqueName: \"kubernetes.io/projected/dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd-kube-api-access-n7ksn\") pod \"marketplace-operator-79b997595-t4mbz\" (UID: \"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.174662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.413181 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wfbxx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.413476 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.601217 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t4mbz"] Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.717622 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.752878 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.910432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-utilities\") pod \"e93c2521-42a0-4944-97c7-448e498294a4\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.910782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m9b7\" (UniqueName: \"kubernetes.io/projected/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-kube-api-access-6m9b7\") pod \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.910817 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrxvr\" (UniqueName: \"kubernetes.io/projected/e93c2521-42a0-4944-97c7-448e498294a4-kube-api-access-wrxvr\") pod \"e93c2521-42a0-4944-97c7-448e498294a4\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.910841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-catalog-content\") pod \"e93c2521-42a0-4944-97c7-448e498294a4\" (UID: \"e93c2521-42a0-4944-97c7-448e498294a4\") " Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.910886 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-catalog-content\") pod \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.910920 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-utilities\") pod \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\" (UID: \"5aaa8ff8-bbd1-4768-9325-285eb8acb01a\") " Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.912137 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-utilities" (OuterVolumeSpecName: "utilities") pod "e93c2521-42a0-4944-97c7-448e498294a4" (UID: "e93c2521-42a0-4944-97c7-448e498294a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.912681 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-utilities" (OuterVolumeSpecName: "utilities") pod "5aaa8ff8-bbd1-4768-9325-285eb8acb01a" (UID: "5aaa8ff8-bbd1-4768-9325-285eb8acb01a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.916393 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93c2521-42a0-4944-97c7-448e498294a4-kube-api-access-wrxvr" (OuterVolumeSpecName: "kube-api-access-wrxvr") pod "e93c2521-42a0-4944-97c7-448e498294a4" (UID: "e93c2521-42a0-4944-97c7-448e498294a4"). InnerVolumeSpecName "kube-api-access-wrxvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.916536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-kube-api-access-6m9b7" (OuterVolumeSpecName: "kube-api-access-6m9b7") pod "5aaa8ff8-bbd1-4768-9325-285eb8acb01a" (UID: "5aaa8ff8-bbd1-4768-9325-285eb8acb01a"). InnerVolumeSpecName "kube-api-access-6m9b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:24 crc kubenswrapper[4809]: I1205 11:15:24.967910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5aaa8ff8-bbd1-4768-9325-285eb8acb01a" (UID: "5aaa8ff8-bbd1-4768-9325-285eb8acb01a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.011846 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m9b7\" (UniqueName: \"kubernetes.io/projected/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-kube-api-access-6m9b7\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.011889 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrxvr\" (UniqueName: \"kubernetes.io/projected/e93c2521-42a0-4944-97c7-448e498294a4-kube-api-access-wrxvr\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.011900 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.011908 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaa8ff8-bbd1-4768-9325-285eb8acb01a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.011916 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.021806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e93c2521-42a0-4944-97c7-448e498294a4" (UID: "e93c2521-42a0-4944-97c7-448e498294a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.112565 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93c2521-42a0-4944-97c7-448e498294a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.133396 4809 generic.go:334] "Generic (PLEG): container finished" podID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerID="cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74" exitCode=0 Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.133501 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9t5j4" event={"ID":"75df67eb-0996-4396-94bb-8ddbce4136ce","Type":"ContainerDied","Data":"cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.135364 4809 generic.go:334] "Generic (PLEG): container finished" podID="e93c2521-42a0-4944-97c7-448e498294a4" containerID="1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92" exitCode=0 Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.135522 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrg7v" event={"ID":"e93c2521-42a0-4944-97c7-448e498294a4","Type":"ContainerDied","Data":"1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.135604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrg7v" event={"ID":"e93c2521-42a0-4944-97c7-448e498294a4","Type":"ContainerDied","Data":"6af6b0fd47016acf44a939ea845477ed3796925630bbba3127cf7b17c26df7ba"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.135739 4809 scope.go:117] "RemoveContainer" containerID="1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.136051 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrg7v" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.139523 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" event={"ID":"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd","Type":"ContainerStarted","Data":"dfaa656981ebe11daccb2dc18fc0462f9b330ef41be2a285984c90706abc2910"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.139569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" event={"ID":"dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd","Type":"ContainerStarted","Data":"9c8e96bf471d3224abe0558077d93b6681d907d9fad5e7c611f24d8816d63de3"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.140016 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.158144 4809 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-t4mbz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.159179 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" podUID="dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.162675 4809 generic.go:334] "Generic (PLEG): container finished" podID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerID="74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de" exitCode=0 Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.162734 4809 scope.go:117] "RemoveContainer" containerID="769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.162752 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xj244" event={"ID":"5aaa8ff8-bbd1-4768-9325-285eb8acb01a","Type":"ContainerDied","Data":"74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.162776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xj244" event={"ID":"5aaa8ff8-bbd1-4768-9325-285eb8acb01a","Type":"ContainerDied","Data":"7cca5619e65eafd6f3e9e1fdeefb4680bb5309e3638a10b4e42a29ae6cf3e58b"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.162850 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xj244" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.173452 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" podStartSLOduration=2.173439186 podStartE2EDuration="2.173439186s" podCreationTimestamp="2025-12-05 11:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:15:25.173257101 +0000 UTC m=+420.564233669" watchObservedRunningTime="2025-12-05 11:15:25.173439186 +0000 UTC m=+420.564415744" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.176279 4809 generic.go:334] "Generic (PLEG): container finished" podID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerID="0c70a1f8ffbd26b6c7010cb1914a3f5ebbddb2881f44cd4598b78da3d444bc60" exitCode=0 Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.176643 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" event={"ID":"8781fe56-ed07-4998-b905-7fd3fc750e42","Type":"ContainerDied","Data":"0c70a1f8ffbd26b6c7010cb1914a3f5ebbddb2881f44cd4598b78da3d444bc60"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.190059 4809 generic.go:334] "Generic (PLEG): container finished" podID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerID="4a4a8a556e52e0d31f19203c359c066f8865ba75e234a413712498d815eb0b00" exitCode=0 Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.190290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcp4f" event={"ID":"29be0f64-8b6e-453e-a8c0-dbe464da2705","Type":"ContainerDied","Data":"4a4a8a556e52e0d31f19203c359c066f8865ba75e234a413712498d815eb0b00"} Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.194167 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jrg7v"] Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.197110 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jrg7v"] Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.213221 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xj244"] Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.221139 4809 scope.go:117] "RemoveContainer" containerID="bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.225353 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xj244"] Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.241006 4809 scope.go:117] "RemoveContainer" containerID="1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92" Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.241300 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92\": container with ID starting with 1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92 not found: ID does not exist" containerID="1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.241331 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92"} err="failed to get container status \"1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92\": rpc error: code = NotFound desc = could not find container \"1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92\": container with ID starting with 1be46def393e4492939fcb9d33cd706dbbed13282923102455b72873ba234c92 not found: ID does not exist" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.241353 4809 scope.go:117] "RemoveContainer" containerID="769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461" Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.241520 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461\": container with ID starting with 769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461 not found: ID does not exist" containerID="769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.241540 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461"} err="failed to get container status \"769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461\": rpc error: code = NotFound desc = could not find container \"769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461\": container with ID starting with 769135e0cfed221d5d1674f9b51db2182a2677a85180359c713ba8fc5c611461 not found: ID does not exist" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.241553 4809 scope.go:117] "RemoveContainer" containerID="bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113" Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.241973 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113\": container with ID starting with bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113 not found: ID does not exist" containerID="bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.241995 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113"} err="failed to get container status \"bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113\": rpc error: code = NotFound desc = could not find container \"bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113\": container with ID starting with bfae0d759f66349b760f5cdd2c480283d89b4ef883411ad2cb1d30ceb6cc4113 not found: ID does not exist" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.242009 4809 scope.go:117] "RemoveContainer" containerID="74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.256282 4809 scope.go:117] "RemoveContainer" containerID="c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.297896 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.306552 4809 scope.go:117] "RemoveContainer" containerID="c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.328498 4809 scope.go:117] "RemoveContainer" containerID="74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de" Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.329252 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de\": container with ID starting with 74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de not found: ID does not exist" containerID="74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.329283 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de"} err="failed to get container status \"74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de\": rpc error: code = NotFound desc = could not find container \"74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de\": container with ID starting with 74f5166e000682df5d6d855d75b311d584d037f784acf58ba103f116d85a87de not found: ID does not exist" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.329310 4809 scope.go:117] "RemoveContainer" containerID="c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51" Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.329700 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51\": container with ID starting with c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51 not found: ID does not exist" containerID="c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.329741 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51"} err="failed to get container status \"c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51\": rpc error: code = NotFound desc = could not find container \"c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51\": container with ID starting with c8aea52f6f6c6a374e7ebe5e9751d05fe17ba09f32005a24830b4094587c4a51 not found: ID does not exist" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.329780 4809 scope.go:117] "RemoveContainer" containerID="c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf" Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.330192 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf\": container with ID starting with c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf not found: ID does not exist" containerID="c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.330221 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf"} err="failed to get container status \"c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf\": rpc error: code = NotFound desc = could not find container \"c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf\": container with ID starting with c564568fe6fba38c4a792bcfec384cf5e1e1132712b6a6207aa44cf309c284cf not found: ID does not exist" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.330239 4809 scope.go:117] "RemoveContainer" containerID="1a746b32726ed8c572c8ecf84ccee922f17e05b980e33563e411bc7a0f460d7e" Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.348791 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74 is running failed: container process not found" containerID="cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.349072 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74 is running failed: container process not found" containerID="cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.349303 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74 is running failed: container process not found" containerID="cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 11:15:25 crc kubenswrapper[4809]: E1205 11:15:25.349338 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-9t5j4" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="registry-server" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.406182 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.412954 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.417781 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hq4g\" (UniqueName: \"kubernetes.io/projected/29be0f64-8b6e-453e-a8c0-dbe464da2705-kube-api-access-9hq4g\") pod \"29be0f64-8b6e-453e-a8c0-dbe464da2705\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.417900 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-catalog-content\") pod \"29be0f64-8b6e-453e-a8c0-dbe464da2705\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.418002 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-utilities\") pod \"29be0f64-8b6e-453e-a8c0-dbe464da2705\" (UID: \"29be0f64-8b6e-453e-a8c0-dbe464da2705\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.419165 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-utilities" (OuterVolumeSpecName: "utilities") pod "29be0f64-8b6e-453e-a8c0-dbe464da2705" (UID: "29be0f64-8b6e-453e-a8c0-dbe464da2705"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.421468 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29be0f64-8b6e-453e-a8c0-dbe464da2705-kube-api-access-9hq4g" (OuterVolumeSpecName: "kube-api-access-9hq4g") pod "29be0f64-8b6e-453e-a8c0-dbe464da2705" (UID: "29be0f64-8b6e-453e-a8c0-dbe464da2705"). InnerVolumeSpecName "kube-api-access-9hq4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.458219 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29be0f64-8b6e-453e-a8c0-dbe464da2705" (UID: "29be0f64-8b6e-453e-a8c0-dbe464da2705"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.518994 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-trusted-ca\") pod \"8781fe56-ed07-4998-b905-7fd3fc750e42\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxjwx\" (UniqueName: \"kubernetes.io/projected/75df67eb-0996-4396-94bb-8ddbce4136ce-kube-api-access-nxjwx\") pod \"75df67eb-0996-4396-94bb-8ddbce4136ce\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-operator-metrics\") pod \"8781fe56-ed07-4998-b905-7fd3fc750e42\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519209 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-utilities\") pod \"75df67eb-0996-4396-94bb-8ddbce4136ce\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519243 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4cn4\" (UniqueName: \"kubernetes.io/projected/8781fe56-ed07-4998-b905-7fd3fc750e42-kube-api-access-l4cn4\") pod \"8781fe56-ed07-4998-b905-7fd3fc750e42\" (UID: \"8781fe56-ed07-4998-b905-7fd3fc750e42\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519289 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-catalog-content\") pod \"75df67eb-0996-4396-94bb-8ddbce4136ce\" (UID: \"75df67eb-0996-4396-94bb-8ddbce4136ce\") " Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519527 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519553 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hq4g\" (UniqueName: \"kubernetes.io/projected/29be0f64-8b6e-453e-a8c0-dbe464da2705-kube-api-access-9hq4g\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.519572 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29be0f64-8b6e-453e-a8c0-dbe464da2705-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.521105 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-utilities" (OuterVolumeSpecName: "utilities") pod "75df67eb-0996-4396-94bb-8ddbce4136ce" (UID: "75df67eb-0996-4396-94bb-8ddbce4136ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.521131 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8781fe56-ed07-4998-b905-7fd3fc750e42" (UID: "8781fe56-ed07-4998-b905-7fd3fc750e42"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.522555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75df67eb-0996-4396-94bb-8ddbce4136ce-kube-api-access-nxjwx" (OuterVolumeSpecName: "kube-api-access-nxjwx") pod "75df67eb-0996-4396-94bb-8ddbce4136ce" (UID: "75df67eb-0996-4396-94bb-8ddbce4136ce"). InnerVolumeSpecName "kube-api-access-nxjwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.522830 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8781fe56-ed07-4998-b905-7fd3fc750e42-kube-api-access-l4cn4" (OuterVolumeSpecName: "kube-api-access-l4cn4") pod "8781fe56-ed07-4998-b905-7fd3fc750e42" (UID: "8781fe56-ed07-4998-b905-7fd3fc750e42"). InnerVolumeSpecName "kube-api-access-l4cn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.524402 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8781fe56-ed07-4998-b905-7fd3fc750e42" (UID: "8781fe56-ed07-4998-b905-7fd3fc750e42"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.564796 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75df67eb-0996-4396-94bb-8ddbce4136ce" (UID: "75df67eb-0996-4396-94bb-8ddbce4136ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.619930 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4cn4\" (UniqueName: \"kubernetes.io/projected/8781fe56-ed07-4998-b905-7fd3fc750e42-kube-api-access-l4cn4\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.619958 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.619970 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.619978 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxjwx\" (UniqueName: \"kubernetes.io/projected/75df67eb-0996-4396-94bb-8ddbce4136ce-kube-api-access-nxjwx\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.619986 4809 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8781fe56-ed07-4998-b905-7fd3fc750e42-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:25 crc kubenswrapper[4809]: I1205 11:15:25.619995 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75df67eb-0996-4396-94bb-8ddbce4136ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.197857 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.197955 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wfbxx" event={"ID":"8781fe56-ed07-4998-b905-7fd3fc750e42","Type":"ContainerDied","Data":"2da6558d17dcb8d098d6a5970114f9ffc3142551d1f3e3aac22bd29d6b99ce28"} Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.207814 4809 scope.go:117] "RemoveContainer" containerID="0c70a1f8ffbd26b6c7010cb1914a3f5ebbddb2881f44cd4598b78da3d444bc60" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.213172 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcp4f" event={"ID":"29be0f64-8b6e-453e-a8c0-dbe464da2705","Type":"ContainerDied","Data":"1e2301b3a09c3dd63861051666265c2ced971e2fd192ce1c5c235aceb8db777e"} Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.213241 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcp4f" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.215299 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wkvnz"] Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.215888 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217396 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217424 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217433 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217447 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217455 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217468 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217476 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217486 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217493 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217507 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217517 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217531 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217539 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217548 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217613 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217681 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217692 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217702 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217710 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217719 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217727 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="extract-content" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217740 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217748 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217761 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217770 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: E1205 11:15:26.217780 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.217788 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="extract-utilities" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.218016 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.218033 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.218043 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.218051 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.218067 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" containerName="marketplace-operator" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.218078 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e93c2521-42a0-4944-97c7-448e498294a4" containerName="registry-server" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.218308 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9t5j4" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.221654 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9t5j4" event={"ID":"75df67eb-0996-4396-94bb-8ddbce4136ce","Type":"ContainerDied","Data":"0fab17bc7202aac0e33e18c4e3718817ee6309197f48955b3c35cd1e4238ee62"} Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.221868 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.226919 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.233015 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-t4mbz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.237339 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wkvnz"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.269713 4809 scope.go:117] "RemoveContainer" containerID="4a4a8a556e52e0d31f19203c359c066f8865ba75e234a413712498d815eb0b00" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.291133 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9t5j4"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.295487 4809 scope.go:117] "RemoveContainer" containerID="20e630fd820bbc96a2e4f10a51fa3cc6705b8dc2bfddbf6e1e3023b24d322fb2" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.295926 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9t5j4"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.304297 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wfbxx"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.308866 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wfbxx"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.318117 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcp4f"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.322478 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcp4f"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.323138 4809 scope.go:117] "RemoveContainer" containerID="9fae6f55d1ebbf8eacf831d0bef3dc0bc19650b742b754a1cd6851a54d352b6a" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.329714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b07cc015-c350-4282-84e3-7bd761097773-catalog-content\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.329818 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b07cc015-c350-4282-84e3-7bd761097773-utilities\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.329853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sknls\" (UniqueName: \"kubernetes.io/projected/b07cc015-c350-4282-84e3-7bd761097773-kube-api-access-sknls\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.335536 4809 scope.go:117] "RemoveContainer" containerID="cea86a4a8ffda7703299354e9ebbb2e79f7a2b7cee3b4ff924e3e3db2e622d74" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.350054 4809 scope.go:117] "RemoveContainer" containerID="f8d97c64c93f70d66aa091259bbd8c120436dbb7a847e37590fabc06667ce778" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.368793 4809 scope.go:117] "RemoveContainer" containerID="4fb36aaccd438d4f57c7e7a19206bfca03603bf8381db1d876e9131355a0d985" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.432212 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b07cc015-c350-4282-84e3-7bd761097773-utilities\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.432277 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sknls\" (UniqueName: \"kubernetes.io/projected/b07cc015-c350-4282-84e3-7bd761097773-kube-api-access-sknls\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.432315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b07cc015-c350-4282-84e3-7bd761097773-catalog-content\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.432940 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b07cc015-c350-4282-84e3-7bd761097773-catalog-content\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.433124 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b07cc015-c350-4282-84e3-7bd761097773-utilities\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.449658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sknls\" (UniqueName: \"kubernetes.io/projected/b07cc015-c350-4282-84e3-7bd761097773-kube-api-access-sknls\") pod \"community-operators-wkvnz\" (UID: \"b07cc015-c350-4282-84e3-7bd761097773\") " pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.578048 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.747100 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wkvnz"] Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.879224 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29be0f64-8b6e-453e-a8c0-dbe464da2705" path="/var/lib/kubelet/pods/29be0f64-8b6e-453e-a8c0-dbe464da2705/volumes" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.880056 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aaa8ff8-bbd1-4768-9325-285eb8acb01a" path="/var/lib/kubelet/pods/5aaa8ff8-bbd1-4768-9325-285eb8acb01a/volumes" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.880864 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75df67eb-0996-4396-94bb-8ddbce4136ce" path="/var/lib/kubelet/pods/75df67eb-0996-4396-94bb-8ddbce4136ce/volumes" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.882257 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8781fe56-ed07-4998-b905-7fd3fc750e42" path="/var/lib/kubelet/pods/8781fe56-ed07-4998-b905-7fd3fc750e42/volumes" Dec 05 11:15:26 crc kubenswrapper[4809]: I1205 11:15:26.882951 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e93c2521-42a0-4944-97c7-448e498294a4" path="/var/lib/kubelet/pods/e93c2521-42a0-4944-97c7-448e498294a4/volumes" Dec 05 11:15:27 crc kubenswrapper[4809]: I1205 11:15:27.233590 4809 generic.go:334] "Generic (PLEG): container finished" podID="b07cc015-c350-4282-84e3-7bd761097773" containerID="a6d5dff1a5d8a31e168bc0f81f3ce19af6139e0798e6d517111e987ebc0f0f92" exitCode=0 Dec 05 11:15:27 crc kubenswrapper[4809]: I1205 11:15:27.233677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wkvnz" event={"ID":"b07cc015-c350-4282-84e3-7bd761097773","Type":"ContainerDied","Data":"a6d5dff1a5d8a31e168bc0f81f3ce19af6139e0798e6d517111e987ebc0f0f92"} Dec 05 11:15:27 crc kubenswrapper[4809]: I1205 11:15:27.233735 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wkvnz" event={"ID":"b07cc015-c350-4282-84e3-7bd761097773","Type":"ContainerStarted","Data":"984b3cb15c2e3910e676d4de36af66d5c5af33c1de8d78175a59d811b12a2e33"} Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.015505 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4pm6h"] Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.018751 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.021217 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.022596 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4pm6h"] Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.062240 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724c514f-9629-482f-aa1d-f750ed364603-utilities\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.062446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq9xr\" (UniqueName: \"kubernetes.io/projected/724c514f-9629-482f-aa1d-f750ed364603-kube-api-access-zq9xr\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.062526 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724c514f-9629-482f-aa1d-f750ed364603-catalog-content\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.163494 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724c514f-9629-482f-aa1d-f750ed364603-utilities\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.163575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq9xr\" (UniqueName: \"kubernetes.io/projected/724c514f-9629-482f-aa1d-f750ed364603-kube-api-access-zq9xr\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.163607 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724c514f-9629-482f-aa1d-f750ed364603-catalog-content\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.164337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724c514f-9629-482f-aa1d-f750ed364603-utilities\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.164342 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724c514f-9629-482f-aa1d-f750ed364603-catalog-content\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.185171 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq9xr\" (UniqueName: \"kubernetes.io/projected/724c514f-9629-482f-aa1d-f750ed364603-kube-api-access-zq9xr\") pod \"redhat-operators-4pm6h\" (UID: \"724c514f-9629-482f-aa1d-f750ed364603\") " pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.242221 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wkvnz" event={"ID":"b07cc015-c350-4282-84e3-7bd761097773","Type":"ContainerStarted","Data":"8bc85c4ac887f1d179039a4dbf0d24e362eddd16669496a0cb56398fc14c6f29"} Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.337870 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.611858 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nk4cv"] Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.612972 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.615002 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.620398 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk4cv"] Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.670402 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-catalog-content\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.670484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8rlf\" (UniqueName: \"kubernetes.io/projected/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-kube-api-access-n8rlf\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.670536 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-utilities\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.762827 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4pm6h"] Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.771161 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-catalog-content\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.771204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8rlf\" (UniqueName: \"kubernetes.io/projected/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-kube-api-access-n8rlf\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.771239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-utilities\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.771752 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-catalog-content\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.771850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-utilities\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.791076 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8rlf\" (UniqueName: \"kubernetes.io/projected/a1b2282b-0f31-46c0-8b85-5e41e5bb1248-kube-api-access-n8rlf\") pod \"certified-operators-nk4cv\" (UID: \"a1b2282b-0f31-46c0-8b85-5e41e5bb1248\") " pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:28 crc kubenswrapper[4809]: I1205 11:15:28.938257 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:29 crc kubenswrapper[4809]: I1205 11:15:29.136076 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk4cv"] Dec 05 11:15:29 crc kubenswrapper[4809]: W1205 11:15:29.143364 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1b2282b_0f31_46c0_8b85_5e41e5bb1248.slice/crio-9683af278a5ce920e3b9a6594239eae5bdfe31b250de716a75cd62f081d7674d WatchSource:0}: Error finding container 9683af278a5ce920e3b9a6594239eae5bdfe31b250de716a75cd62f081d7674d: Status 404 returned error can't find the container with id 9683af278a5ce920e3b9a6594239eae5bdfe31b250de716a75cd62f081d7674d Dec 05 11:15:29 crc kubenswrapper[4809]: I1205 11:15:29.248367 4809 generic.go:334] "Generic (PLEG): container finished" podID="724c514f-9629-482f-aa1d-f750ed364603" containerID="224cdfd6fa7093d54a5d8bd07e30b1f46fc0a2e8f6fd1eb18ebc8458b77e9f35" exitCode=0 Dec 05 11:15:29 crc kubenswrapper[4809]: I1205 11:15:29.248453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pm6h" event={"ID":"724c514f-9629-482f-aa1d-f750ed364603","Type":"ContainerDied","Data":"224cdfd6fa7093d54a5d8bd07e30b1f46fc0a2e8f6fd1eb18ebc8458b77e9f35"} Dec 05 11:15:29 crc kubenswrapper[4809]: I1205 11:15:29.248502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pm6h" event={"ID":"724c514f-9629-482f-aa1d-f750ed364603","Type":"ContainerStarted","Data":"75610fbc322843b602edd90b5724186d1c5db1599fb39dd449b8cb0904dfafdf"} Dec 05 11:15:29 crc kubenswrapper[4809]: I1205 11:15:29.249928 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4cv" event={"ID":"a1b2282b-0f31-46c0-8b85-5e41e5bb1248","Type":"ContainerStarted","Data":"9683af278a5ce920e3b9a6594239eae5bdfe31b250de716a75cd62f081d7674d"} Dec 05 11:15:29 crc kubenswrapper[4809]: I1205 11:15:29.252217 4809 generic.go:334] "Generic (PLEG): container finished" podID="b07cc015-c350-4282-84e3-7bd761097773" containerID="8bc85c4ac887f1d179039a4dbf0d24e362eddd16669496a0cb56398fc14c6f29" exitCode=0 Dec 05 11:15:29 crc kubenswrapper[4809]: I1205 11:15:29.252255 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wkvnz" event={"ID":"b07cc015-c350-4282-84e3-7bd761097773","Type":"ContainerDied","Data":"8bc85c4ac887f1d179039a4dbf0d24e362eddd16669496a0cb56398fc14c6f29"} Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.264439 4809 generic.go:334] "Generic (PLEG): container finished" podID="a1b2282b-0f31-46c0-8b85-5e41e5bb1248" containerID="f6da039a6fea0f861f1b7275a27a3f9d192d685b20a696cbf0557d131bce1920" exitCode=0 Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.264761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4cv" event={"ID":"a1b2282b-0f31-46c0-8b85-5e41e5bb1248","Type":"ContainerDied","Data":"f6da039a6fea0f861f1b7275a27a3f9d192d685b20a696cbf0557d131bce1920"} Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.411845 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qspwd"] Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.413203 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.415513 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.435339 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qspwd"] Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.495040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-utilities\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.495117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wndbg\" (UniqueName: \"kubernetes.io/projected/1299366a-d263-4417-aef3-eaa199c32e72-kube-api-access-wndbg\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.495343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-catalog-content\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.596353 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-catalog-content\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.596401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-utilities\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.596425 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wndbg\" (UniqueName: \"kubernetes.io/projected/1299366a-d263-4417-aef3-eaa199c32e72-kube-api-access-wndbg\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.596928 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-utilities\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.596934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-catalog-content\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.618330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wndbg\" (UniqueName: \"kubernetes.io/projected/1299366a-d263-4417-aef3-eaa199c32e72-kube-api-access-wndbg\") pod \"redhat-marketplace-qspwd\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:30 crc kubenswrapper[4809]: I1205 11:15:30.730276 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:31 crc kubenswrapper[4809]: I1205 11:15:31.195530 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qspwd"] Dec 05 11:15:31 crc kubenswrapper[4809]: W1205 11:15:31.202112 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1299366a_d263_4417_aef3_eaa199c32e72.slice/crio-1aeb51f4f79fade7cebb86c435a59de22221a814fbc2f431819f25e52079b36b WatchSource:0}: Error finding container 1aeb51f4f79fade7cebb86c435a59de22221a814fbc2f431819f25e52079b36b: Status 404 returned error can't find the container with id 1aeb51f4f79fade7cebb86c435a59de22221a814fbc2f431819f25e52079b36b Dec 05 11:15:31 crc kubenswrapper[4809]: I1205 11:15:31.273317 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4cv" event={"ID":"a1b2282b-0f31-46c0-8b85-5e41e5bb1248","Type":"ContainerStarted","Data":"56aae40a0c30c32be4d6249cda4b58103c320520e88c90c2e1bc2be908dd883b"} Dec 05 11:15:31 crc kubenswrapper[4809]: I1205 11:15:31.275805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wkvnz" event={"ID":"b07cc015-c350-4282-84e3-7bd761097773","Type":"ContainerStarted","Data":"c18f74fa8ba6434e00fb77e0e629c3182b82c9f68bbb568a4ac0a3da9edca526"} Dec 05 11:15:31 crc kubenswrapper[4809]: I1205 11:15:31.276904 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qspwd" event={"ID":"1299366a-d263-4417-aef3-eaa199c32e72","Type":"ContainerStarted","Data":"1aeb51f4f79fade7cebb86c435a59de22221a814fbc2f431819f25e52079b36b"} Dec 05 11:15:31 crc kubenswrapper[4809]: I1205 11:15:31.278406 4809 generic.go:334] "Generic (PLEG): container finished" podID="724c514f-9629-482f-aa1d-f750ed364603" containerID="443126844a1d2ecb9e911e38b9da07b96e24a614f759975784f948fb8c718d3c" exitCode=0 Dec 05 11:15:31 crc kubenswrapper[4809]: I1205 11:15:31.278435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pm6h" event={"ID":"724c514f-9629-482f-aa1d-f750ed364603","Type":"ContainerDied","Data":"443126844a1d2ecb9e911e38b9da07b96e24a614f759975784f948fb8c718d3c"} Dec 05 11:15:31 crc kubenswrapper[4809]: I1205 11:15:31.309175 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wkvnz" podStartSLOduration=2.532407025 podStartE2EDuration="5.309150688s" podCreationTimestamp="2025-12-05 11:15:26 +0000 UTC" firstStartedPulling="2025-12-05 11:15:27.235599424 +0000 UTC m=+422.626576002" lastFinishedPulling="2025-12-05 11:15:30.012343097 +0000 UTC m=+425.403319665" observedRunningTime="2025-12-05 11:15:31.302916289 +0000 UTC m=+426.693892857" watchObservedRunningTime="2025-12-05 11:15:31.309150688 +0000 UTC m=+426.700127246" Dec 05 11:15:32 crc kubenswrapper[4809]: I1205 11:15:32.287150 4809 generic.go:334] "Generic (PLEG): container finished" podID="a1b2282b-0f31-46c0-8b85-5e41e5bb1248" containerID="56aae40a0c30c32be4d6249cda4b58103c320520e88c90c2e1bc2be908dd883b" exitCode=0 Dec 05 11:15:32 crc kubenswrapper[4809]: I1205 11:15:32.287223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4cv" event={"ID":"a1b2282b-0f31-46c0-8b85-5e41e5bb1248","Type":"ContainerDied","Data":"56aae40a0c30c32be4d6249cda4b58103c320520e88c90c2e1bc2be908dd883b"} Dec 05 11:15:32 crc kubenswrapper[4809]: I1205 11:15:32.289468 4809 generic.go:334] "Generic (PLEG): container finished" podID="1299366a-d263-4417-aef3-eaa199c32e72" containerID="550c1b347633229b7d6ab14d55033e90af99ce4d5ce5e3e7aa046b8c18c31f8f" exitCode=0 Dec 05 11:15:32 crc kubenswrapper[4809]: I1205 11:15:32.289535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qspwd" event={"ID":"1299366a-d263-4417-aef3-eaa199c32e72","Type":"ContainerDied","Data":"550c1b347633229b7d6ab14d55033e90af99ce4d5ce5e3e7aa046b8c18c31f8f"} Dec 05 11:15:32 crc kubenswrapper[4809]: I1205 11:15:32.294610 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4pm6h" event={"ID":"724c514f-9629-482f-aa1d-f750ed364603","Type":"ContainerStarted","Data":"465c351c18c2641ab468786bbbbf40de9672087e8331ca08107d40a5d23a2ce6"} Dec 05 11:15:32 crc kubenswrapper[4809]: I1205 11:15:32.340056 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4pm6h" podStartSLOduration=2.694861489 podStartE2EDuration="5.340038233s" podCreationTimestamp="2025-12-05 11:15:27 +0000 UTC" firstStartedPulling="2025-12-05 11:15:29.249770682 +0000 UTC m=+424.640747240" lastFinishedPulling="2025-12-05 11:15:31.894947426 +0000 UTC m=+427.285923984" observedRunningTime="2025-12-05 11:15:32.335355839 +0000 UTC m=+427.726332397" watchObservedRunningTime="2025-12-05 11:15:32.340038233 +0000 UTC m=+427.731014791" Dec 05 11:15:33 crc kubenswrapper[4809]: I1205 11:15:33.300685 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk4cv" event={"ID":"a1b2282b-0f31-46c0-8b85-5e41e5bb1248","Type":"ContainerStarted","Data":"4e154fabe1a33063f5974a35ad1638ec0d0c8d17f1439518ce0500698c4afc5b"} Dec 05 11:15:33 crc kubenswrapper[4809]: I1205 11:15:33.302300 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qspwd" event={"ID":"1299366a-d263-4417-aef3-eaa199c32e72","Type":"ContainerStarted","Data":"7871fe02a77d8fa46af144fc6b2460c9c1123dea09d90a4d0addd66b1d9e3b8f"} Dec 05 11:15:33 crc kubenswrapper[4809]: I1205 11:15:33.319578 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nk4cv" podStartSLOduration=2.674600351 podStartE2EDuration="5.319559562s" podCreationTimestamp="2025-12-05 11:15:28 +0000 UTC" firstStartedPulling="2025-12-05 11:15:30.267488712 +0000 UTC m=+425.658465270" lastFinishedPulling="2025-12-05 11:15:32.912447923 +0000 UTC m=+428.303424481" observedRunningTime="2025-12-05 11:15:33.318533919 +0000 UTC m=+428.709510477" watchObservedRunningTime="2025-12-05 11:15:33.319559562 +0000 UTC m=+428.710536120" Dec 05 11:15:34 crc kubenswrapper[4809]: I1205 11:15:34.309452 4809 generic.go:334] "Generic (PLEG): container finished" podID="1299366a-d263-4417-aef3-eaa199c32e72" containerID="7871fe02a77d8fa46af144fc6b2460c9c1123dea09d90a4d0addd66b1d9e3b8f" exitCode=0 Dec 05 11:15:34 crc kubenswrapper[4809]: I1205 11:15:34.309648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qspwd" event={"ID":"1299366a-d263-4417-aef3-eaa199c32e72","Type":"ContainerDied","Data":"7871fe02a77d8fa46af144fc6b2460c9c1123dea09d90a4d0addd66b1d9e3b8f"} Dec 05 11:15:36 crc kubenswrapper[4809]: I1205 11:15:36.323187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qspwd" event={"ID":"1299366a-d263-4417-aef3-eaa199c32e72","Type":"ContainerStarted","Data":"5dcb8d720a19e31dd679c03e814aff742aae95e3b123a38e0cd881ceb3211818"} Dec 05 11:15:36 crc kubenswrapper[4809]: I1205 11:15:36.578658 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:36 crc kubenswrapper[4809]: I1205 11:15:36.578925 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:36 crc kubenswrapper[4809]: I1205 11:15:36.623364 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:37 crc kubenswrapper[4809]: I1205 11:15:37.346555 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qspwd" podStartSLOduration=4.554189614 podStartE2EDuration="7.346536335s" podCreationTimestamp="2025-12-05 11:15:30 +0000 UTC" firstStartedPulling="2025-12-05 11:15:32.290558098 +0000 UTC m=+427.681534656" lastFinishedPulling="2025-12-05 11:15:35.082904819 +0000 UTC m=+430.473881377" observedRunningTime="2025-12-05 11:15:37.344565571 +0000 UTC m=+432.735542159" watchObservedRunningTime="2025-12-05 11:15:37.346536335 +0000 UTC m=+432.737512893" Dec 05 11:15:37 crc kubenswrapper[4809]: I1205 11:15:37.367750 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wkvnz" Dec 05 11:15:38 crc kubenswrapper[4809]: I1205 11:15:38.339145 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:38 crc kubenswrapper[4809]: I1205 11:15:38.339197 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:38 crc kubenswrapper[4809]: I1205 11:15:38.380959 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:38 crc kubenswrapper[4809]: I1205 11:15:38.938593 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:38 crc kubenswrapper[4809]: I1205 11:15:38.938703 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:38 crc kubenswrapper[4809]: I1205 11:15:38.983724 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:39 crc kubenswrapper[4809]: I1205 11:15:39.372123 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4pm6h" Dec 05 11:15:39 crc kubenswrapper[4809]: I1205 11:15:39.379568 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nk4cv" Dec 05 11:15:40 crc kubenswrapper[4809]: I1205 11:15:40.730776 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:40 crc kubenswrapper[4809]: I1205 11:15:40.731156 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:40 crc kubenswrapper[4809]: I1205 11:15:40.777052 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:41 crc kubenswrapper[4809]: I1205 11:15:41.385289 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:15:44 crc kubenswrapper[4809]: I1205 11:15:44.047891 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:15:44 crc kubenswrapper[4809]: I1205 11:15:44.047996 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:15:44 crc kubenswrapper[4809]: I1205 11:15:44.048078 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:15:44 crc kubenswrapper[4809]: I1205 11:15:44.051776 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7b326672db1ccbd336f3353e4dddf0e3ea107025106a857e28fa99fc3ca732a"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:15:44 crc kubenswrapper[4809]: I1205 11:15:44.052022 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://e7b326672db1ccbd336f3353e4dddf0e3ea107025106a857e28fa99fc3ca732a" gracePeriod=600 Dec 05 11:15:46 crc kubenswrapper[4809]: I1205 11:15:46.369556 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="e7b326672db1ccbd336f3353e4dddf0e3ea107025106a857e28fa99fc3ca732a" exitCode=0 Dec 05 11:15:46 crc kubenswrapper[4809]: I1205 11:15:46.369600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"e7b326672db1ccbd336f3353e4dddf0e3ea107025106a857e28fa99fc3ca732a"} Dec 05 11:15:46 crc kubenswrapper[4809]: I1205 11:15:46.370141 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"269ccb7d83be79c33038b2361609297104c7f320bc76f310c4d65273c96c6d58"} Dec 05 11:15:46 crc kubenswrapper[4809]: I1205 11:15:46.370164 4809 scope.go:117] "RemoveContainer" containerID="fbc4694d4d60440e8bc2162b2fbaaea0ecae60e4c43787e503887ce004c87ed4" Dec 05 11:18:14 crc kubenswrapper[4809]: I1205 11:18:14.047268 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:18:14 crc kubenswrapper[4809]: I1205 11:18:14.047812 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:18:44 crc kubenswrapper[4809]: I1205 11:18:44.046675 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:18:44 crc kubenswrapper[4809]: I1205 11:18:44.047181 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.046742 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.047313 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.047366 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.048103 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"269ccb7d83be79c33038b2361609297104c7f320bc76f310c4d65273c96c6d58"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.048228 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://269ccb7d83be79c33038b2361609297104c7f320bc76f310c4d65273c96c6d58" gracePeriod=600 Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.626093 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="269ccb7d83be79c33038b2361609297104c7f320bc76f310c4d65273c96c6d58" exitCode=0 Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.626144 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"269ccb7d83be79c33038b2361609297104c7f320bc76f310c4d65273c96c6d58"} Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.626559 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"95559463623c60efa0b7c2c68dd34e50145b75a4e23fb8a25ef08e1782d3a372"} Dec 05 11:19:14 crc kubenswrapper[4809]: I1205 11:19:14.626582 4809 scope.go:117] "RemoveContainer" containerID="e7b326672db1ccbd336f3353e4dddf0e3ea107025106a857e28fa99fc3ca732a" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.205867 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zvm2f"] Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.207023 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.227304 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zvm2f"] Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d3724f6b-3c50-4601-8e3f-2e84639ce762-registry-certificates\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mflm\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-kube-api-access-7mflm\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273566 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d3724f6b-3c50-4601-8e3f-2e84639ce762-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-bound-sa-token\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273662 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d3724f6b-3c50-4601-8e3f-2e84639ce762-trusted-ca\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273685 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d3724f6b-3c50-4601-8e3f-2e84639ce762-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-registry-tls\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.273752 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.289970 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.374908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mflm\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-kube-api-access-7mflm\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.374951 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d3724f6b-3c50-4601-8e3f-2e84639ce762-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.375006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-bound-sa-token\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.375054 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d3724f6b-3c50-4601-8e3f-2e84639ce762-trusted-ca\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.375116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d3724f6b-3c50-4601-8e3f-2e84639ce762-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.375979 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d3724f6b-3c50-4601-8e3f-2e84639ce762-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.376409 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d3724f6b-3c50-4601-8e3f-2e84639ce762-trusted-ca\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.376477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-registry-tls\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.376874 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d3724f6b-3c50-4601-8e3f-2e84639ce762-registry-certificates\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.388072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-registry-tls\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.390347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d3724f6b-3c50-4601-8e3f-2e84639ce762-registry-certificates\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.390747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mflm\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-kube-api-access-7mflm\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.397099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d3724f6b-3c50-4601-8e3f-2e84639ce762-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.397912 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d3724f6b-3c50-4601-8e3f-2e84639ce762-bound-sa-token\") pod \"image-registry-66df7c8f76-zvm2f\" (UID: \"d3724f6b-3c50-4601-8e3f-2e84639ce762\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.530816 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.735149 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zvm2f"] Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.982384 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" event={"ID":"d3724f6b-3c50-4601-8e3f-2e84639ce762","Type":"ContainerStarted","Data":"cf467759681c3533e69bc67732a8e376a04acdd9cb2009cd88c6e6c4068dab9e"} Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.982829 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:11 crc kubenswrapper[4809]: I1205 11:20:11.982852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" event={"ID":"d3724f6b-3c50-4601-8e3f-2e84639ce762","Type":"ContainerStarted","Data":"a12c3ed6924baee6664957c4d5f466f9bb81c2757bc5a2ae7b56a9dff1ab2234"} Dec 05 11:20:31 crc kubenswrapper[4809]: I1205 11:20:31.536929 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" Dec 05 11:20:31 crc kubenswrapper[4809]: I1205 11:20:31.557440 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-zvm2f" podStartSLOduration=20.557423852 podStartE2EDuration="20.557423852s" podCreationTimestamp="2025-12-05 11:20:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:20:12.014743081 +0000 UTC m=+707.405719669" watchObservedRunningTime="2025-12-05 11:20:31.557423852 +0000 UTC m=+726.948400410" Dec 05 11:20:31 crc kubenswrapper[4809]: I1205 11:20:31.610458 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx7gs"] Dec 05 11:20:56 crc kubenswrapper[4809]: I1205 11:20:56.644036 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" podUID="1de639b4-f614-41b9-812b-a7bfea3ccc90" containerName="registry" containerID="cri-o://824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf" gracePeriod=30 Dec 05 11:20:56 crc kubenswrapper[4809]: I1205 11:20:56.997478 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082450 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1de639b4-f614-41b9-812b-a7bfea3ccc90-ca-trust-extracted\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082558 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1de639b4-f614-41b9-812b-a7bfea3ccc90-installation-pull-secrets\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082616 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-bound-sa-token\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082662 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-certificates\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082692 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89rhj\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-kube-api-access-89rhj\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-tls\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.082748 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-trusted-ca\") pod \"1de639b4-f614-41b9-812b-a7bfea3ccc90\" (UID: \"1de639b4-f614-41b9-812b-a7bfea3ccc90\") " Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.086580 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.087942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.088489 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-kube-api-access-89rhj" (OuterVolumeSpecName: "kube-api-access-89rhj") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "kube-api-access-89rhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.088809 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.090035 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.090741 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1de639b4-f614-41b9-812b-a7bfea3ccc90-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.093048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.098455 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1de639b4-f614-41b9-812b-a7bfea3ccc90-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1de639b4-f614-41b9-812b-a7bfea3ccc90" (UID: "1de639b4-f614-41b9-812b-a7bfea3ccc90"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.183938 4809 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.183976 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.183986 4809 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1de639b4-f614-41b9-812b-a7bfea3ccc90-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.183996 4809 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1de639b4-f614-41b9-812b-a7bfea3ccc90-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.184008 4809 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.184016 4809 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1de639b4-f614-41b9-812b-a7bfea3ccc90-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.184024 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89rhj\" (UniqueName: \"kubernetes.io/projected/1de639b4-f614-41b9-812b-a7bfea3ccc90-kube-api-access-89rhj\") on node \"crc\" DevicePath \"\"" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.344209 4809 generic.go:334] "Generic (PLEG): container finished" podID="1de639b4-f614-41b9-812b-a7bfea3ccc90" containerID="824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf" exitCode=0 Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.344251 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" event={"ID":"1de639b4-f614-41b9-812b-a7bfea3ccc90","Type":"ContainerDied","Data":"824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf"} Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.344277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" event={"ID":"1de639b4-f614-41b9-812b-a7bfea3ccc90","Type":"ContainerDied","Data":"8587c79bbc94e25ef6c22c3a9eab8bc8faf7580d3dbc2bfc1a9afb80f966a492"} Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.344297 4809 scope.go:117] "RemoveContainer" containerID="824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.344294 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mx7gs" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.360179 4809 scope.go:117] "RemoveContainer" containerID="824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf" Dec 05 11:20:57 crc kubenswrapper[4809]: E1205 11:20:57.361310 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf\": container with ID starting with 824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf not found: ID does not exist" containerID="824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.361338 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf"} err="failed to get container status \"824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf\": rpc error: code = NotFound desc = could not find container \"824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf\": container with ID starting with 824aa4c67ad6f3f7edba81c26aabee15058e1062ed9e472aba20b898df93c1cf not found: ID does not exist" Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.372825 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx7gs"] Dec 05 11:20:57 crc kubenswrapper[4809]: I1205 11:20:57.376240 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mx7gs"] Dec 05 11:20:58 crc kubenswrapper[4809]: I1205 11:20:58.881351 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de639b4-f614-41b9-812b-a7bfea3ccc90" path="/var/lib/kubelet/pods/1de639b4-f614-41b9-812b-a7bfea3ccc90/volumes" Dec 05 11:21:06 crc kubenswrapper[4809]: I1205 11:21:06.658431 4809 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 11:21:14 crc kubenswrapper[4809]: I1205 11:21:14.046453 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:21:14 crc kubenswrapper[4809]: I1205 11:21:14.046976 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:21:44 crc kubenswrapper[4809]: I1205 11:21:44.046339 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:21:44 crc kubenswrapper[4809]: I1205 11:21:44.047252 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.046977 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.047688 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.047776 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.048708 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95559463623c60efa0b7c2c68dd34e50145b75a4e23fb8a25ef08e1782d3a372"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.048828 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://95559463623c60efa0b7c2c68dd34e50145b75a4e23fb8a25ef08e1782d3a372" gracePeriod=600 Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.866499 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="95559463623c60efa0b7c2c68dd34e50145b75a4e23fb8a25ef08e1782d3a372" exitCode=0 Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.866566 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"95559463623c60efa0b7c2c68dd34e50145b75a4e23fb8a25ef08e1782d3a372"} Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.866914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"bd6b669d62c0db11d4d46d14f9c5c9d47f99d27bb3324e58211f411bd95dcbe7"} Dec 05 11:22:14 crc kubenswrapper[4809]: I1205 11:22:14.866949 4809 scope.go:117] "RemoveContainer" containerID="269ccb7d83be79c33038b2361609297104c7f320bc76f310c4d65273c96c6d58" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.592333 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cdvk8"] Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.593858 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-controller" containerID="cri-o://1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.594080 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="northd" containerID="cri-o://3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.594283 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-node" containerID="cri-o://90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.594348 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-acl-logging" containerID="cri-o://91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.594366 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="nbdb" containerID="cri-o://95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.594341 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="sbdb" containerID="cri-o://30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.594447 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.646190 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" containerID="cri-o://e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" gracePeriod=30 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.878176 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/3.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.880741 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovn-acl-logging/0.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.881304 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovn-controller/0.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.881766 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.916518 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/2.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.916983 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/1.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.917016 4809 generic.go:334] "Generic (PLEG): container finished" podID="238d6f1b-f3b4-4b33-a33c-10a0c99e8178" containerID="5861e92534d270984c5ea325ee01014773b774019a667ed88e3490f52b54c167" exitCode=2 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.917077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerDied","Data":"5861e92534d270984c5ea325ee01014773b774019a667ed88e3490f52b54c167"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.917129 4809 scope.go:117] "RemoveContainer" containerID="65177ae6827b12456691816d3cc67f19bb54a880c19c4c4caeaefb0196d81cbf" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.917875 4809 scope.go:117] "RemoveContainer" containerID="5861e92534d270984c5ea325ee01014773b774019a667ed88e3490f52b54c167" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.922183 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovnkube-controller/3.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.930483 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovn-acl-logging/0.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.931126 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cdvk8_df22f1c5-aea5-4717-b2b0-3d6a577ea53f/ovn-controller/0.log" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932107 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" exitCode=0 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932264 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" exitCode=0 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932276 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" exitCode=0 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932284 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" exitCode=0 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932296 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" exitCode=0 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932304 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" exitCode=0 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932312 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" exitCode=143 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932320 4809 generic.go:334] "Generic (PLEG): container finished" podID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerID="1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" exitCode=143 Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932340 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932366 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932378 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932388 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932416 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932426 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932431 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932436 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932441 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932446 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932451 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932456 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932460 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932465 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932479 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932487 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932492 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932497 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932502 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932507 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932513 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932517 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932524 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932529 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932543 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932549 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932554 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932559 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932564 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932569 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932575 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932579 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932584 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932589 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" event={"ID":"df22f1c5-aea5-4717-b2b0-3d6a577ea53f","Type":"ContainerDied","Data":"69838862339e86709318f40582d66ac51e9d319f9ca09bc0a40b49034ea0cc87"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932604 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932611 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932616 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932621 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932626 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932646 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932651 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932657 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932662 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932620 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cdvk8" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.932667 4809 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958038 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kt4zh"] Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958357 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="northd" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958384 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="northd" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958400 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de639b4-f614-41b9-812b-a7bfea3ccc90" containerName="registry" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958410 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de639b4-f614-41b9-812b-a7bfea3ccc90" containerName="registry" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958427 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958435 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958446 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958454 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958465 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958474 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958484 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958493 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958507 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-acl-logging" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958516 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-acl-logging" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958527 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958535 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958545 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-node" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958553 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-node" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958563 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="nbdb" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958572 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="nbdb" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958585 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958594 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958609 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kubecfg-setup" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958617 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kubecfg-setup" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.958672 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="sbdb" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958681 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="sbdb" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958801 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de639b4-f614-41b9-812b-a7bfea3ccc90" containerName="registry" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958814 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-acl-logging" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958824 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958837 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-node" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958848 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovn-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958859 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958870 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="nbdb" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958882 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="sbdb" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958891 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="northd" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958899 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.958910 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: E1205 11:22:22.959040 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.959054 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.959193 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.959501 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" containerName="ovnkube-controller" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.963495 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:22 crc kubenswrapper[4809]: I1205 11:22:22.970671 4809 scope.go:117] "RemoveContainer" containerID="e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-openvswitch\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003844 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-kubelet\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003868 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-systemd\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003893 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003933 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-ovn\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003958 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-etc-openvswitch\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003994 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-systemd-units\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-var-lib-openvswitch\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004046 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovn-node-metrics-cert\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004073 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-script-lib\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003438 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004111 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-netd\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003954 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.003989 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004028 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004087 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-log-socket" (OuterVolumeSpecName: "log-socket") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004130 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-log-socket\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-bin\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004244 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-netns\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004280 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-ovn-kubernetes\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4kb5\" (UniqueName: \"kubernetes.io/projected/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-kube-api-access-j4kb5\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004355 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-env-overrides\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004388 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-slash\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004450 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-node-log\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004485 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-config\") pod \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\" (UID: \"df22f1c5-aea5-4717-b2b0-3d6a577ea53f\") " Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004772 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004881 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-node-log" (OuterVolumeSpecName: "node-log") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.005255 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.004845 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.005346 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.005786 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-slash" (OuterVolumeSpecName: "host-slash") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.005970 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.005976 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006099 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006117 4809 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-log-socket\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006126 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006135 4809 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006146 4809 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-node-log\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006154 4809 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006163 4809 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006172 4809 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006181 4809 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006189 4809 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006199 4809 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.006207 4809 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.010289 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.010291 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-kube-api-access-j4kb5" (OuterVolumeSpecName: "kube-api-access-j4kb5") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "kube-api-access-j4kb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.020408 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "df22f1c5-aea5-4717-b2b0-3d6a577ea53f" (UID: "df22f1c5-aea5-4717-b2b0-3d6a577ea53f"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.023748 4809 scope.go:117] "RemoveContainer" containerID="30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.038916 4809 scope.go:117] "RemoveContainer" containerID="95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.051356 4809 scope.go:117] "RemoveContainer" containerID="3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.064981 4809 scope.go:117] "RemoveContainer" containerID="d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.078785 4809 scope.go:117] "RemoveContainer" containerID="90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.092349 4809 scope.go:117] "RemoveContainer" containerID="91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.106816 4809 scope.go:117] "RemoveContainer" containerID="1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107356 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovn-node-metrics-cert\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107407 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-var-lib-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-cni-bin\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-run-netns\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107646 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-cni-netd\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107729 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107766 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-slash\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107823 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-node-log\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-run-ovn-kubernetes\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107898 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-systemd-units\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.107921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovnkube-config\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovnkube-script-lib\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-log-socket\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-env-overrides\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77h2l\" (UniqueName: \"kubernetes.io/projected/6c091f2d-dbbe-4d80-af63-705da27c1dfe-kube-api-access-77h2l\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108127 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-ovn\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108152 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-kubelet\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-systemd\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108217 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-etc-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108301 4809 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108321 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108334 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108349 4809 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108362 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4kb5\" (UniqueName: \"kubernetes.io/projected/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-kube-api-access-j4kb5\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108374 4809 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108386 4809 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-host-slash\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.108398 4809 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df22f1c5-aea5-4717-b2b0-3d6a577ea53f-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.123350 4809 scope.go:117] "RemoveContainer" containerID="89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.136178 4809 scope.go:117] "RemoveContainer" containerID="e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.136726 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": container with ID starting with e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e not found: ID does not exist" containerID="e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.136772 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} err="failed to get container status \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": rpc error: code = NotFound desc = could not find container \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": container with ID starting with e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.136866 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.137306 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": container with ID starting with e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b not found: ID does not exist" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.137347 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} err="failed to get container status \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": rpc error: code = NotFound desc = could not find container \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": container with ID starting with e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.137385 4809 scope.go:117] "RemoveContainer" containerID="30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.137833 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": container with ID starting with 30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a not found: ID does not exist" containerID="30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.137877 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} err="failed to get container status \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": rpc error: code = NotFound desc = could not find container \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": container with ID starting with 30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.137896 4809 scope.go:117] "RemoveContainer" containerID="95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.138190 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": container with ID starting with 95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4 not found: ID does not exist" containerID="95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.138221 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} err="failed to get container status \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": rpc error: code = NotFound desc = could not find container \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": container with ID starting with 95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.138241 4809 scope.go:117] "RemoveContainer" containerID="3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.138555 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": container with ID starting with 3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549 not found: ID does not exist" containerID="3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.138593 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} err="failed to get container status \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": rpc error: code = NotFound desc = could not find container \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": container with ID starting with 3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.138606 4809 scope.go:117] "RemoveContainer" containerID="d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.138895 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": container with ID starting with d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb not found: ID does not exist" containerID="d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.138913 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} err="failed to get container status \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": rpc error: code = NotFound desc = could not find container \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": container with ID starting with d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.138942 4809 scope.go:117] "RemoveContainer" containerID="90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.139209 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": container with ID starting with 90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5 not found: ID does not exist" containerID="90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.139232 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} err="failed to get container status \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": rpc error: code = NotFound desc = could not find container \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": container with ID starting with 90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.139249 4809 scope.go:117] "RemoveContainer" containerID="91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.139506 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": container with ID starting with 91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f not found: ID does not exist" containerID="91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.139534 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} err="failed to get container status \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": rpc error: code = NotFound desc = could not find container \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": container with ID starting with 91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.139550 4809 scope.go:117] "RemoveContainer" containerID="1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.139897 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": container with ID starting with 1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6 not found: ID does not exist" containerID="1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.139931 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} err="failed to get container status \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": rpc error: code = NotFound desc = could not find container \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": container with ID starting with 1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.139945 4809 scope.go:117] "RemoveContainer" containerID="89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f" Dec 05 11:22:23 crc kubenswrapper[4809]: E1205 11:22:23.140349 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": container with ID starting with 89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f not found: ID does not exist" containerID="89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.140391 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} err="failed to get container status \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": rpc error: code = NotFound desc = could not find container \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": container with ID starting with 89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.140406 4809 scope.go:117] "RemoveContainer" containerID="e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.140741 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} err="failed to get container status \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": rpc error: code = NotFound desc = could not find container \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": container with ID starting with e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.140778 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.141193 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} err="failed to get container status \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": rpc error: code = NotFound desc = could not find container \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": container with ID starting with e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.141216 4809 scope.go:117] "RemoveContainer" containerID="30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.141584 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} err="failed to get container status \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": rpc error: code = NotFound desc = could not find container \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": container with ID starting with 30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.141607 4809 scope.go:117] "RemoveContainer" containerID="95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.141861 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} err="failed to get container status \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": rpc error: code = NotFound desc = could not find container \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": container with ID starting with 95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.141881 4809 scope.go:117] "RemoveContainer" containerID="3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142180 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} err="failed to get container status \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": rpc error: code = NotFound desc = could not find container \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": container with ID starting with 3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142198 4809 scope.go:117] "RemoveContainer" containerID="d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142428 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} err="failed to get container status \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": rpc error: code = NotFound desc = could not find container \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": container with ID starting with d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142449 4809 scope.go:117] "RemoveContainer" containerID="90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142670 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} err="failed to get container status \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": rpc error: code = NotFound desc = could not find container \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": container with ID starting with 90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142704 4809 scope.go:117] "RemoveContainer" containerID="91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142901 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} err="failed to get container status \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": rpc error: code = NotFound desc = could not find container \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": container with ID starting with 91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.142920 4809 scope.go:117] "RemoveContainer" containerID="1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.143192 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} err="failed to get container status \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": rpc error: code = NotFound desc = could not find container \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": container with ID starting with 1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.143212 4809 scope.go:117] "RemoveContainer" containerID="89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.143506 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} err="failed to get container status \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": rpc error: code = NotFound desc = could not find container \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": container with ID starting with 89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.143528 4809 scope.go:117] "RemoveContainer" containerID="e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.143867 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} err="failed to get container status \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": rpc error: code = NotFound desc = could not find container \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": container with ID starting with e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.143928 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.144227 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} err="failed to get container status \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": rpc error: code = NotFound desc = could not find container \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": container with ID starting with e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.144246 4809 scope.go:117] "RemoveContainer" containerID="30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.144524 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} err="failed to get container status \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": rpc error: code = NotFound desc = could not find container \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": container with ID starting with 30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.144544 4809 scope.go:117] "RemoveContainer" containerID="95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.144756 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} err="failed to get container status \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": rpc error: code = NotFound desc = could not find container \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": container with ID starting with 95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.144772 4809 scope.go:117] "RemoveContainer" containerID="3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.145117 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} err="failed to get container status \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": rpc error: code = NotFound desc = could not find container \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": container with ID starting with 3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.145145 4809 scope.go:117] "RemoveContainer" containerID="d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.145485 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} err="failed to get container status \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": rpc error: code = NotFound desc = could not find container \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": container with ID starting with d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.145516 4809 scope.go:117] "RemoveContainer" containerID="90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.145788 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} err="failed to get container status \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": rpc error: code = NotFound desc = could not find container \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": container with ID starting with 90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.145810 4809 scope.go:117] "RemoveContainer" containerID="91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.146010 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} err="failed to get container status \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": rpc error: code = NotFound desc = could not find container \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": container with ID starting with 91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.146027 4809 scope.go:117] "RemoveContainer" containerID="1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.146304 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} err="failed to get container status \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": rpc error: code = NotFound desc = could not find container \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": container with ID starting with 1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.146342 4809 scope.go:117] "RemoveContainer" containerID="89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.146656 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} err="failed to get container status \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": rpc error: code = NotFound desc = could not find container \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": container with ID starting with 89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.146678 4809 scope.go:117] "RemoveContainer" containerID="e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.147690 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e"} err="failed to get container status \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": rpc error: code = NotFound desc = could not find container \"e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e\": container with ID starting with e5cddbc84d5678047adc24840235ee0ee8c2e705bb3ee73bd4f4ebf2e4aff87e not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.147788 4809 scope.go:117] "RemoveContainer" containerID="e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.148302 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b"} err="failed to get container status \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": rpc error: code = NotFound desc = could not find container \"e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b\": container with ID starting with e8caf510330cb6ab482b4444d479715146dec52d39d138fa8af02aed5ca0c90b not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.148322 4809 scope.go:117] "RemoveContainer" containerID="30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.148718 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a"} err="failed to get container status \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": rpc error: code = NotFound desc = could not find container \"30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a\": container with ID starting with 30860769a44238a6b1739dfa1f0cc7b687b9f4cb278d8e7b3cf407d26b5a754a not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.148744 4809 scope.go:117] "RemoveContainer" containerID="95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.149345 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4"} err="failed to get container status \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": rpc error: code = NotFound desc = could not find container \"95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4\": container with ID starting with 95bf13fbe8b4d9e13f62516ed238eaaf6bba7c0c314c413b8bc4f99f30976ad4 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.149361 4809 scope.go:117] "RemoveContainer" containerID="3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.149644 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549"} err="failed to get container status \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": rpc error: code = NotFound desc = could not find container \"3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549\": container with ID starting with 3b29ecdefc9b6d1b0ec8db7748695cd8e3dc744ca0be0d55ccd5cbe89b51d549 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.149671 4809 scope.go:117] "RemoveContainer" containerID="d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.149931 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb"} err="failed to get container status \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": rpc error: code = NotFound desc = could not find container \"d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb\": container with ID starting with d140e444f62281f3fbe16d5bddf09b6e648c2afb28fa2ef91350e2e9d47768cb not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.149949 4809 scope.go:117] "RemoveContainer" containerID="90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.150170 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5"} err="failed to get container status \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": rpc error: code = NotFound desc = could not find container \"90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5\": container with ID starting with 90cad978711d0fa89e40e28816a1ec8670198833badc1c95d98d6a6524c3b5d5 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.150193 4809 scope.go:117] "RemoveContainer" containerID="91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.150394 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f"} err="failed to get container status \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": rpc error: code = NotFound desc = could not find container \"91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f\": container with ID starting with 91d3ebe40b88a54e56a6021fc42c640259d0d4e4622a3ee677723238dee7182f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.150419 4809 scope.go:117] "RemoveContainer" containerID="1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.150682 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6"} err="failed to get container status \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": rpc error: code = NotFound desc = could not find container \"1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6\": container with ID starting with 1452a4c614145b441d6ed66520799435834b8ddc5343b4b8e4e2bdf17e4e07d6 not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.150739 4809 scope.go:117] "RemoveContainer" containerID="89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.151253 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f"} err="failed to get container status \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": rpc error: code = NotFound desc = could not find container \"89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f\": container with ID starting with 89e5e39a1b96984afe72368fe79a8a960757b60b525b2543e98902bac8eba43f not found: ID does not exist" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209734 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovn-node-metrics-cert\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-var-lib-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-cni-bin\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-run-netns\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-cni-netd\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209861 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209879 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-run-netns\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209929 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-slash\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209950 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209906 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-slash\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209961 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-cni-netd\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.209979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-node-log\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-run-ovn-kubernetes\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210020 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-systemd-units\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210057 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-run-ovn-kubernetes\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210085 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-node-log\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-systemd-units\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210015 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-cni-bin\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210037 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovnkube-config\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-var-lib-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210201 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovnkube-script-lib\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-log-socket\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210275 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-env-overrides\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210310 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77h2l\" (UniqueName: \"kubernetes.io/projected/6c091f2d-dbbe-4d80-af63-705da27c1dfe-kube-api-access-77h2l\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210331 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-ovn\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-systemd\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210382 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-kubelet\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-etc-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-systemd\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210418 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-log-socket\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-run-ovn\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210451 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-host-kubelet\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6c091f2d-dbbe-4d80-af63-705da27c1dfe-etc-openvswitch\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210776 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovnkube-config\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.210876 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-env-overrides\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.211441 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovnkube-script-lib\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.213352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6c091f2d-dbbe-4d80-af63-705da27c1dfe-ovn-node-metrics-cert\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.226789 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77h2l\" (UniqueName: \"kubernetes.io/projected/6c091f2d-dbbe-4d80-af63-705da27c1dfe-kube-api-access-77h2l\") pod \"ovnkube-node-kt4zh\" (UID: \"6c091f2d-dbbe-4d80-af63-705da27c1dfe\") " pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.271618 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cdvk8"] Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.271682 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cdvk8"] Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.290184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:23 crc kubenswrapper[4809]: W1205 11:22:23.310275 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c091f2d_dbbe_4d80_af63_705da27c1dfe.slice/crio-4008856d78fd07ae8aa283606987af7941f749ed3d57cfcd5634abfd646a30e0 WatchSource:0}: Error finding container 4008856d78fd07ae8aa283606987af7941f749ed3d57cfcd5634abfd646a30e0: Status 404 returned error can't find the container with id 4008856d78fd07ae8aa283606987af7941f749ed3d57cfcd5634abfd646a30e0 Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.944802 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n79nb_238d6f1b-f3b4-4b33-a33c-10a0c99e8178/kube-multus/2.log" Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.945420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n79nb" event={"ID":"238d6f1b-f3b4-4b33-a33c-10a0c99e8178","Type":"ContainerStarted","Data":"8cac881c0f96033848bb6861a535adfa10319db358c36e13c198a3133c242c80"} Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.951520 4809 generic.go:334] "Generic (PLEG): container finished" podID="6c091f2d-dbbe-4d80-af63-705da27c1dfe" containerID="4053f1abfd4ad3aac215e8c8d8b645b0666338b21807174ca8e64e423003dae8" exitCode=0 Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.951602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerDied","Data":"4053f1abfd4ad3aac215e8c8d8b645b0666338b21807174ca8e64e423003dae8"} Dec 05 11:22:23 crc kubenswrapper[4809]: I1205 11:22:23.951654 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"4008856d78fd07ae8aa283606987af7941f749ed3d57cfcd5634abfd646a30e0"} Dec 05 11:22:24 crc kubenswrapper[4809]: I1205 11:22:24.890892 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df22f1c5-aea5-4717-b2b0-3d6a577ea53f" path="/var/lib/kubelet/pods/df22f1c5-aea5-4717-b2b0-3d6a577ea53f/volumes" Dec 05 11:22:24 crc kubenswrapper[4809]: I1205 11:22:24.961194 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"95aeefac536a9161ac20750711356bf509bdf851038fa6e94fa9067862e88e2c"} Dec 05 11:22:24 crc kubenswrapper[4809]: I1205 11:22:24.961237 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"57c614b96e622d6777ee0267ec2e031435ac682373acd9fc524131e1986435f6"} Dec 05 11:22:24 crc kubenswrapper[4809]: I1205 11:22:24.961247 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"4861b0a75942b226585e9bb648b93d296aa94c759dd8fe7c494349a892ea08e9"} Dec 05 11:22:24 crc kubenswrapper[4809]: I1205 11:22:24.961256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"a849de551ced92323ea4615828421885d960d836071019ca688dfed2376205d8"} Dec 05 11:22:24 crc kubenswrapper[4809]: I1205 11:22:24.961264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"9abc9922aa8d1bfeefda510c5b8317813453919a873b10b6a2352abe5271fb42"} Dec 05 11:22:24 crc kubenswrapper[4809]: I1205 11:22:24.961272 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"c9994557bdc9c1c3ec03fffbeb1e782ff0d1a64e08163085a670f16bb1e081e9"} Dec 05 11:22:27 crc kubenswrapper[4809]: I1205 11:22:27.987214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"ed3b4b7b8b88a141e1745dfbbfabd7266d7b509004c91c34a978781a5ebc4a99"} Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.153347 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-xxhbm"] Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.154357 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.156790 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.157745 4809 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-lngh5" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.157784 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.158139 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.290521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99v9k\" (UniqueName: \"kubernetes.io/projected/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-kube-api-access-99v9k\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.290598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-node-mnt\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.290701 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-crc-storage\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.391544 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-node-mnt\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.391699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-crc-storage\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.391854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99v9k\" (UniqueName: \"kubernetes.io/projected/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-kube-api-access-99v9k\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.391961 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-node-mnt\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.393586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-crc-storage\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.423340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99v9k\" (UniqueName: \"kubernetes.io/projected/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-kube-api-access-99v9k\") pod \"crc-storage-crc-xxhbm\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: I1205 11:22:29.472804 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: E1205 11:22:29.502252 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(0e59c31a9c664bc90fc978485f4bac71a94edf57fe517bd8c9ba140fd298f9c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 11:22:29 crc kubenswrapper[4809]: E1205 11:22:29.502354 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(0e59c31a9c664bc90fc978485f4bac71a94edf57fe517bd8c9ba140fd298f9c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: E1205 11:22:29.502388 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(0e59c31a9c664bc90fc978485f4bac71a94edf57fe517bd8c9ba140fd298f9c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:29 crc kubenswrapper[4809]: E1205 11:22:29.502461 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-xxhbm_crc-storage(6f4be64e-7da9-41d2-95b6-bf38e7cc254e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-xxhbm_crc-storage(6f4be64e-7da9-41d2-95b6-bf38e7cc254e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(0e59c31a9c664bc90fc978485f4bac71a94edf57fe517bd8c9ba140fd298f9c4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-xxhbm" podUID="6f4be64e-7da9-41d2-95b6-bf38e7cc254e" Dec 05 11:22:30 crc kubenswrapper[4809]: I1205 11:22:30.010450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" event={"ID":"6c091f2d-dbbe-4d80-af63-705da27c1dfe","Type":"ContainerStarted","Data":"851f701f1bb9016233169e08a5eb7cfb7bd027861483b74c3c2d80ad81fc458b"} Dec 05 11:22:30 crc kubenswrapper[4809]: I1205 11:22:30.774006 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-xxhbm"] Dec 05 11:22:30 crc kubenswrapper[4809]: I1205 11:22:30.774121 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:30 crc kubenswrapper[4809]: I1205 11:22:30.774757 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:30 crc kubenswrapper[4809]: E1205 11:22:30.800383 4809 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(1a1befd5be1f63bd2d978068eeeeae15f314622aea622966ce665ce79f3c627d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 11:22:30 crc kubenswrapper[4809]: E1205 11:22:30.800504 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(1a1befd5be1f63bd2d978068eeeeae15f314622aea622966ce665ce79f3c627d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:30 crc kubenswrapper[4809]: E1205 11:22:30.800538 4809 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(1a1befd5be1f63bd2d978068eeeeae15f314622aea622966ce665ce79f3c627d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:30 crc kubenswrapper[4809]: E1205 11:22:30.800645 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-xxhbm_crc-storage(6f4be64e-7da9-41d2-95b6-bf38e7cc254e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-xxhbm_crc-storage(6f4be64e-7da9-41d2-95b6-bf38e7cc254e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xxhbm_crc-storage_6f4be64e-7da9-41d2-95b6-bf38e7cc254e_0(1a1befd5be1f63bd2d978068eeeeae15f314622aea622966ce665ce79f3c627d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-xxhbm" podUID="6f4be64e-7da9-41d2-95b6-bf38e7cc254e" Dec 05 11:22:31 crc kubenswrapper[4809]: I1205 11:22:31.017703 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:31 crc kubenswrapper[4809]: I1205 11:22:31.017749 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:31 crc kubenswrapper[4809]: I1205 11:22:31.049166 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:31 crc kubenswrapper[4809]: I1205 11:22:31.052756 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" podStartSLOduration=9.052741749 podStartE2EDuration="9.052741749s" podCreationTimestamp="2025-12-05 11:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:22:31.049939172 +0000 UTC m=+846.440915730" watchObservedRunningTime="2025-12-05 11:22:31.052741749 +0000 UTC m=+846.443718307" Dec 05 11:22:32 crc kubenswrapper[4809]: I1205 11:22:32.023919 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:32 crc kubenswrapper[4809]: I1205 11:22:32.057051 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:34 crc kubenswrapper[4809]: I1205 11:22:34.069654 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" Dec 05 11:22:42 crc kubenswrapper[4809]: I1205 11:22:42.871868 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:42 crc kubenswrapper[4809]: I1205 11:22:42.873374 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:43 crc kubenswrapper[4809]: I1205 11:22:43.350732 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-xxhbm"] Dec 05 11:22:43 crc kubenswrapper[4809]: I1205 11:22:43.360790 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:22:44 crc kubenswrapper[4809]: I1205 11:22:44.106105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xxhbm" event={"ID":"6f4be64e-7da9-41d2-95b6-bf38e7cc254e","Type":"ContainerStarted","Data":"66a22a61be2e6d7bfad946c75da0e9b3eb9dffc375c77c14df76dbd0efd0015e"} Dec 05 11:22:45 crc kubenswrapper[4809]: I1205 11:22:45.118801 4809 generic.go:334] "Generic (PLEG): container finished" podID="6f4be64e-7da9-41d2-95b6-bf38e7cc254e" containerID="eb7bddf66b3cfad78ed11ced857ef43a6d395bcb8aed6b3ee9b02775b51d267b" exitCode=0 Dec 05 11:22:45 crc kubenswrapper[4809]: I1205 11:22:45.118904 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xxhbm" event={"ID":"6f4be64e-7da9-41d2-95b6-bf38e7cc254e","Type":"ContainerDied","Data":"eb7bddf66b3cfad78ed11ced857ef43a6d395bcb8aed6b3ee9b02775b51d267b"} Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.403756 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.535543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99v9k\" (UniqueName: \"kubernetes.io/projected/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-kube-api-access-99v9k\") pod \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.535694 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-node-mnt\") pod \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.535738 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-crc-storage\") pod \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\" (UID: \"6f4be64e-7da9-41d2-95b6-bf38e7cc254e\") " Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.535910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "6f4be64e-7da9-41d2-95b6-bf38e7cc254e" (UID: "6f4be64e-7da9-41d2-95b6-bf38e7cc254e"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.536112 4809 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.540687 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-kube-api-access-99v9k" (OuterVolumeSpecName: "kube-api-access-99v9k") pod "6f4be64e-7da9-41d2-95b6-bf38e7cc254e" (UID: "6f4be64e-7da9-41d2-95b6-bf38e7cc254e"). InnerVolumeSpecName "kube-api-access-99v9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.557534 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "6f4be64e-7da9-41d2-95b6-bf38e7cc254e" (UID: "6f4be64e-7da9-41d2-95b6-bf38e7cc254e"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.637710 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99v9k\" (UniqueName: \"kubernetes.io/projected/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-kube-api-access-99v9k\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:46 crc kubenswrapper[4809]: I1205 11:22:46.637748 4809 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6f4be64e-7da9-41d2-95b6-bf38e7cc254e-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 05 11:22:47 crc kubenswrapper[4809]: I1205 11:22:47.133887 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xxhbm" event={"ID":"6f4be64e-7da9-41d2-95b6-bf38e7cc254e","Type":"ContainerDied","Data":"66a22a61be2e6d7bfad946c75da0e9b3eb9dffc375c77c14df76dbd0efd0015e"} Dec 05 11:22:47 crc kubenswrapper[4809]: I1205 11:22:47.133941 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66a22a61be2e6d7bfad946c75da0e9b3eb9dffc375c77c14df76dbd0efd0015e" Dec 05 11:22:47 crc kubenswrapper[4809]: I1205 11:22:47.134039 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xxhbm" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.152212 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf"] Dec 05 11:22:55 crc kubenswrapper[4809]: E1205 11:22:55.152823 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f4be64e-7da9-41d2-95b6-bf38e7cc254e" containerName="storage" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.152834 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f4be64e-7da9-41d2-95b6-bf38e7cc254e" containerName="storage" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.152924 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f4be64e-7da9-41d2-95b6-bf38e7cc254e" containerName="storage" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.153543 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.159322 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.168483 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf"] Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.228119 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.228197 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.228525 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvkbv\" (UniqueName: \"kubernetes.io/projected/cd91db76-778c-4208-83f3-9234765c7b8d-kube-api-access-kvkbv\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.330553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.330625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.330830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvkbv\" (UniqueName: \"kubernetes.io/projected/cd91db76-778c-4208-83f3-9234765c7b8d-kube-api-access-kvkbv\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.331818 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.332094 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.351009 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvkbv\" (UniqueName: \"kubernetes.io/projected/cd91db76-778c-4208-83f3-9234765c7b8d-kube-api-access-kvkbv\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.472883 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:22:55 crc kubenswrapper[4809]: I1205 11:22:55.703663 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf"] Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.183685 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd91db76-778c-4208-83f3-9234765c7b8d" containerID="b54792ce9c6784db6a3d1a4f9bda1abc10117eccf071b7ada4fa322b81f31d7b" exitCode=0 Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.183835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" event={"ID":"cd91db76-778c-4208-83f3-9234765c7b8d","Type":"ContainerDied","Data":"b54792ce9c6784db6a3d1a4f9bda1abc10117eccf071b7ada4fa322b81f31d7b"} Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.184828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" event={"ID":"cd91db76-778c-4208-83f3-9234765c7b8d","Type":"ContainerStarted","Data":"e2f702ee176182a9edab6168ad2db94c28edb63769a8ecd989bbf66ac0577c58"} Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.897399 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jlk9c"] Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.899397 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.904567 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jlk9c"] Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.951188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-catalog-content\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.951291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-utilities\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:56 crc kubenswrapper[4809]: I1205 11:22:56.951362 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44brg\" (UniqueName: \"kubernetes.io/projected/6ac37414-e8ca-4039-aa51-762842fb6bb6-kube-api-access-44brg\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.053103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-catalog-content\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.053197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-utilities\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.053254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44brg\" (UniqueName: \"kubernetes.io/projected/6ac37414-e8ca-4039-aa51-762842fb6bb6-kube-api-access-44brg\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.054000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-catalog-content\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.054035 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-utilities\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.080419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44brg\" (UniqueName: \"kubernetes.io/projected/6ac37414-e8ca-4039-aa51-762842fb6bb6-kube-api-access-44brg\") pod \"redhat-operators-jlk9c\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.228990 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:22:57 crc kubenswrapper[4809]: W1205 11:22:57.423396 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ac37414_e8ca_4039_aa51_762842fb6bb6.slice/crio-a9c76a519491366292ec03df7c5295719f7175da52bb057d4a645e27b110a333 WatchSource:0}: Error finding container a9c76a519491366292ec03df7c5295719f7175da52bb057d4a645e27b110a333: Status 404 returned error can't find the container with id a9c76a519491366292ec03df7c5295719f7175da52bb057d4a645e27b110a333 Dec 05 11:22:57 crc kubenswrapper[4809]: I1205 11:22:57.438691 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jlk9c"] Dec 05 11:22:58 crc kubenswrapper[4809]: I1205 11:22:58.197570 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd91db76-778c-4208-83f3-9234765c7b8d" containerID="c943dad49f6a12a495dc7be7f745c3c960089eced1c0031177656aa7d66fe054" exitCode=0 Dec 05 11:22:58 crc kubenswrapper[4809]: I1205 11:22:58.197673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" event={"ID":"cd91db76-778c-4208-83f3-9234765c7b8d","Type":"ContainerDied","Data":"c943dad49f6a12a495dc7be7f745c3c960089eced1c0031177656aa7d66fe054"} Dec 05 11:22:58 crc kubenswrapper[4809]: I1205 11:22:58.202135 4809 generic.go:334] "Generic (PLEG): container finished" podID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerID="6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d" exitCode=0 Dec 05 11:22:58 crc kubenswrapper[4809]: I1205 11:22:58.202220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlk9c" event={"ID":"6ac37414-e8ca-4039-aa51-762842fb6bb6","Type":"ContainerDied","Data":"6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d"} Dec 05 11:22:58 crc kubenswrapper[4809]: I1205 11:22:58.202484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlk9c" event={"ID":"6ac37414-e8ca-4039-aa51-762842fb6bb6","Type":"ContainerStarted","Data":"a9c76a519491366292ec03df7c5295719f7175da52bb057d4a645e27b110a333"} Dec 05 11:22:59 crc kubenswrapper[4809]: I1205 11:22:59.210312 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd91db76-778c-4208-83f3-9234765c7b8d" containerID="f6df5689610aef042c292a26950d993d00d9e107dfd7d6478f58faca0f8b5357" exitCode=0 Dec 05 11:22:59 crc kubenswrapper[4809]: I1205 11:22:59.210419 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" event={"ID":"cd91db76-778c-4208-83f3-9234765c7b8d","Type":"ContainerDied","Data":"f6df5689610aef042c292a26950d993d00d9e107dfd7d6478f58faca0f8b5357"} Dec 05 11:22:59 crc kubenswrapper[4809]: I1205 11:22:59.213584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlk9c" event={"ID":"6ac37414-e8ca-4039-aa51-762842fb6bb6","Type":"ContainerStarted","Data":"4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2"} Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.222450 4809 generic.go:334] "Generic (PLEG): container finished" podID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerID="4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2" exitCode=0 Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.222513 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlk9c" event={"ID":"6ac37414-e8ca-4039-aa51-762842fb6bb6","Type":"ContainerDied","Data":"4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2"} Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.492507 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.595694 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-util\") pod \"cd91db76-778c-4208-83f3-9234765c7b8d\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.595860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvkbv\" (UniqueName: \"kubernetes.io/projected/cd91db76-778c-4208-83f3-9234765c7b8d-kube-api-access-kvkbv\") pod \"cd91db76-778c-4208-83f3-9234765c7b8d\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.595990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-bundle\") pod \"cd91db76-778c-4208-83f3-9234765c7b8d\" (UID: \"cd91db76-778c-4208-83f3-9234765c7b8d\") " Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.596476 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-bundle" (OuterVolumeSpecName: "bundle") pod "cd91db76-778c-4208-83f3-9234765c7b8d" (UID: "cd91db76-778c-4208-83f3-9234765c7b8d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.601206 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd91db76-778c-4208-83f3-9234765c7b8d-kube-api-access-kvkbv" (OuterVolumeSpecName: "kube-api-access-kvkbv") pod "cd91db76-778c-4208-83f3-9234765c7b8d" (UID: "cd91db76-778c-4208-83f3-9234765c7b8d"). InnerVolumeSpecName "kube-api-access-kvkbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.609096 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-util" (OuterVolumeSpecName: "util") pod "cd91db76-778c-4208-83f3-9234765c7b8d" (UID: "cd91db76-778c-4208-83f3-9234765c7b8d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.697858 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.697912 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd91db76-778c-4208-83f3-9234765c7b8d-util\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:00 crc kubenswrapper[4809]: I1205 11:23:00.697937 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvkbv\" (UniqueName: \"kubernetes.io/projected/cd91db76-778c-4208-83f3-9234765c7b8d-kube-api-access-kvkbv\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:01 crc kubenswrapper[4809]: I1205 11:23:01.233293 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" event={"ID":"cd91db76-778c-4208-83f3-9234765c7b8d","Type":"ContainerDied","Data":"e2f702ee176182a9edab6168ad2db94c28edb63769a8ecd989bbf66ac0577c58"} Dec 05 11:23:01 crc kubenswrapper[4809]: I1205 11:23:01.233759 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2f702ee176182a9edab6168ad2db94c28edb63769a8ecd989bbf66ac0577c58" Dec 05 11:23:01 crc kubenswrapper[4809]: I1205 11:23:01.233361 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf" Dec 05 11:23:01 crc kubenswrapper[4809]: I1205 11:23:01.236389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlk9c" event={"ID":"6ac37414-e8ca-4039-aa51-762842fb6bb6","Type":"ContainerStarted","Data":"20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297"} Dec 05 11:23:01 crc kubenswrapper[4809]: I1205 11:23:01.261679 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jlk9c" podStartSLOduration=2.828826152 podStartE2EDuration="5.261610088s" podCreationTimestamp="2025-12-05 11:22:56 +0000 UTC" firstStartedPulling="2025-12-05 11:22:58.203388095 +0000 UTC m=+873.594364653" lastFinishedPulling="2025-12-05 11:23:00.636172021 +0000 UTC m=+876.027148589" observedRunningTime="2025-12-05 11:23:01.254042192 +0000 UTC m=+876.645018840" watchObservedRunningTime="2025-12-05 11:23:01.261610088 +0000 UTC m=+876.652586726" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.398707 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh"] Dec 05 11:23:02 crc kubenswrapper[4809]: E1205 11:23:02.398981 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd91db76-778c-4208-83f3-9234765c7b8d" containerName="extract" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.398997 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd91db76-778c-4208-83f3-9234765c7b8d" containerName="extract" Dec 05 11:23:02 crc kubenswrapper[4809]: E1205 11:23:02.399014 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd91db76-778c-4208-83f3-9234765c7b8d" containerName="pull" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.399022 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd91db76-778c-4208-83f3-9234765c7b8d" containerName="pull" Dec 05 11:23:02 crc kubenswrapper[4809]: E1205 11:23:02.399036 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd91db76-778c-4208-83f3-9234765c7b8d" containerName="util" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.399043 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd91db76-778c-4208-83f3-9234765c7b8d" containerName="util" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.399168 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd91db76-778c-4208-83f3-9234765c7b8d" containerName="extract" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.399600 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.401217 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-wpwtw" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.401402 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.406108 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh"] Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.417975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vrcd\" (UniqueName: \"kubernetes.io/projected/bf3943ce-3694-4815-8d63-6f74dc7c3a9b-kube-api-access-4vrcd\") pod \"nmstate-operator-5b5b58f5c8-xtkjh\" (UID: \"bf3943ce-3694-4815-8d63-6f74dc7c3a9b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.420825 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.519550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vrcd\" (UniqueName: \"kubernetes.io/projected/bf3943ce-3694-4815-8d63-6f74dc7c3a9b-kube-api-access-4vrcd\") pod \"nmstate-operator-5b5b58f5c8-xtkjh\" (UID: \"bf3943ce-3694-4815-8d63-6f74dc7c3a9b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.537076 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vrcd\" (UniqueName: \"kubernetes.io/projected/bf3943ce-3694-4815-8d63-6f74dc7c3a9b-kube-api-access-4vrcd\") pod \"nmstate-operator-5b5b58f5c8-xtkjh\" (UID: \"bf3943ce-3694-4815-8d63-6f74dc7c3a9b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.713157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" Dec 05 11:23:02 crc kubenswrapper[4809]: I1205 11:23:02.928258 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh"] Dec 05 11:23:02 crc kubenswrapper[4809]: W1205 11:23:02.931903 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf3943ce_3694_4815_8d63_6f74dc7c3a9b.slice/crio-88ae9e7e9e660f9e087515f8e4a546c40f241a5d79fe3a8b0281f0b8f483c302 WatchSource:0}: Error finding container 88ae9e7e9e660f9e087515f8e4a546c40f241a5d79fe3a8b0281f0b8f483c302: Status 404 returned error can't find the container with id 88ae9e7e9e660f9e087515f8e4a546c40f241a5d79fe3a8b0281f0b8f483c302 Dec 05 11:23:03 crc kubenswrapper[4809]: I1205 11:23:03.247834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" event={"ID":"bf3943ce-3694-4815-8d63-6f74dc7c3a9b","Type":"ContainerStarted","Data":"88ae9e7e9e660f9e087515f8e4a546c40f241a5d79fe3a8b0281f0b8f483c302"} Dec 05 11:23:06 crc kubenswrapper[4809]: I1205 11:23:06.271223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" event={"ID":"bf3943ce-3694-4815-8d63-6f74dc7c3a9b","Type":"ContainerStarted","Data":"7a0401cc1a0bc3089309891944d75b37638ecacc81b53f6479afaec6908ec07f"} Dec 05 11:23:06 crc kubenswrapper[4809]: I1205 11:23:06.295055 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xtkjh" podStartSLOduration=1.132104163 podStartE2EDuration="4.29501798s" podCreationTimestamp="2025-12-05 11:23:02 +0000 UTC" firstStartedPulling="2025-12-05 11:23:02.933903796 +0000 UTC m=+878.324880354" lastFinishedPulling="2025-12-05 11:23:06.096817613 +0000 UTC m=+881.487794171" observedRunningTime="2025-12-05 11:23:06.284863694 +0000 UTC m=+881.675840282" watchObservedRunningTime="2025-12-05 11:23:06.29501798 +0000 UTC m=+881.685994578" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.229604 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.231175 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.283141 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.285226 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.286778 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jz4r6" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.288440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.291115 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.317729 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.318601 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.323684 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-8ql2q"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.324742 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.332242 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.336008 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.353820 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.390544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq457\" (UniqueName: \"kubernetes.io/projected/f1b67fb0-2d57-4480-9fe5-8862894ce201-kube-api-access-vq457\") pod \"nmstate-webhook-5f6d4c5ccb-pmtjt\" (UID: \"f1b67fb0-2d57-4480-9fe5-8862894ce201\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.390589 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-nmstate-lock\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.390610 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-dbus-socket\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.390653 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mww7v\" (UniqueName: \"kubernetes.io/projected/94e36ccd-7287-4326-bf5e-864dd43c7701-kube-api-access-mww7v\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.390678 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f1b67fb0-2d57-4480-9fe5-8862894ce201-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pmtjt\" (UID: \"f1b67fb0-2d57-4480-9fe5-8862894ce201\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.390718 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6plks\" (UniqueName: \"kubernetes.io/projected/f9d9930b-30f9-4e72-9ebe-f273ab71b519-kube-api-access-6plks\") pod \"nmstate-metrics-7f946cbc9-4wrgc\" (UID: \"f9d9930b-30f9-4e72-9ebe-f273ab71b519\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.390747 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-ovs-socket\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.431023 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.431610 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.435136 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.435280 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.435378 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-j4gvs" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.442792 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492008 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6plks\" (UniqueName: \"kubernetes.io/projected/f9d9930b-30f9-4e72-9ebe-f273ab71b519-kube-api-access-6plks\") pod \"nmstate-metrics-7f946cbc9-4wrgc\" (UID: \"f9d9930b-30f9-4e72-9ebe-f273ab71b519\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492074 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-ovs-socket\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492101 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/701bb0f0-123a-4b4c-b302-c43597314c22-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq457\" (UniqueName: \"kubernetes.io/projected/f1b67fb0-2d57-4480-9fe5-8862894ce201-kube-api-access-vq457\") pod \"nmstate-webhook-5f6d4c5ccb-pmtjt\" (UID: \"f1b67fb0-2d57-4480-9fe5-8862894ce201\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492154 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-nmstate-lock\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-dbus-socket\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-ovs-socket\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mww7v\" (UniqueName: \"kubernetes.io/projected/94e36ccd-7287-4326-bf5e-864dd43c7701-kube-api-access-mww7v\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492240 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqswl\" (UniqueName: \"kubernetes.io/projected/701bb0f0-123a-4b4c-b302-c43597314c22-kube-api-access-wqswl\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492259 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/701bb0f0-123a-4b4c-b302-c43597314c22-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492296 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f1b67fb0-2d57-4480-9fe5-8862894ce201-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pmtjt\" (UID: \"f1b67fb0-2d57-4480-9fe5-8862894ce201\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-nmstate-lock\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.492737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/94e36ccd-7287-4326-bf5e-864dd43c7701-dbus-socket\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.500338 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f1b67fb0-2d57-4480-9fe5-8862894ce201-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pmtjt\" (UID: \"f1b67fb0-2d57-4480-9fe5-8862894ce201\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.505800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq457\" (UniqueName: \"kubernetes.io/projected/f1b67fb0-2d57-4480-9fe5-8862894ce201-kube-api-access-vq457\") pod \"nmstate-webhook-5f6d4c5ccb-pmtjt\" (UID: \"f1b67fb0-2d57-4480-9fe5-8862894ce201\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.508138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mww7v\" (UniqueName: \"kubernetes.io/projected/94e36ccd-7287-4326-bf5e-864dd43c7701-kube-api-access-mww7v\") pod \"nmstate-handler-8ql2q\" (UID: \"94e36ccd-7287-4326-bf5e-864dd43c7701\") " pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.508723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6plks\" (UniqueName: \"kubernetes.io/projected/f9d9930b-30f9-4e72-9ebe-f273ab71b519-kube-api-access-6plks\") pod \"nmstate-metrics-7f946cbc9-4wrgc\" (UID: \"f9d9930b-30f9-4e72-9ebe-f273ab71b519\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.593189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/701bb0f0-123a-4b4c-b302-c43597314c22-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.593290 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqswl\" (UniqueName: \"kubernetes.io/projected/701bb0f0-123a-4b4c-b302-c43597314c22-kube-api-access-wqswl\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.593317 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/701bb0f0-123a-4b4c-b302-c43597314c22-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: E1205 11:23:07.593469 4809 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 05 11:23:07 crc kubenswrapper[4809]: E1205 11:23:07.593528 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/701bb0f0-123a-4b4c-b302-c43597314c22-plugin-serving-cert podName:701bb0f0-123a-4b4c-b302-c43597314c22 nodeName:}" failed. No retries permitted until 2025-12-05 11:23:08.093508769 +0000 UTC m=+883.484485327 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/701bb0f0-123a-4b4c-b302-c43597314c22-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-cqk7j" (UID: "701bb0f0-123a-4b4c-b302-c43597314c22") : secret "plugin-serving-cert" not found Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.594949 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/701bb0f0-123a-4b4c-b302-c43597314c22-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.603134 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.610528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqswl\" (UniqueName: \"kubernetes.io/projected/701bb0f0-123a-4b4c-b302-c43597314c22-kube-api-access-wqswl\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.640969 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.641066 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6dc6c4d949-gnkpx"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.641903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.657069 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.660790 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6dc6c4d949-gnkpx"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.694829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-console-config\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.695165 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e9846351-2ac5-42ae-9332-c97459b1fa06-console-oauth-config\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.695238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e9846351-2ac5-42ae-9332-c97459b1fa06-console-serving-cert\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.695295 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf64q\" (UniqueName: \"kubernetes.io/projected/e9846351-2ac5-42ae-9332-c97459b1fa06-kube-api-access-kf64q\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.695317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-trusted-ca-bundle\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.695345 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-service-ca\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.695376 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-oauth-serving-cert\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.796816 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-trusted-ca-bundle\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.796856 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-service-ca\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.796900 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-oauth-serving-cert\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.796940 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-console-config\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.796974 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e9846351-2ac5-42ae-9332-c97459b1fa06-console-oauth-config\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.797051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e9846351-2ac5-42ae-9332-c97459b1fa06-console-serving-cert\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.797093 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf64q\" (UniqueName: \"kubernetes.io/projected/e9846351-2ac5-42ae-9332-c97459b1fa06-kube-api-access-kf64q\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.798302 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-service-ca\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.799075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-oauth-serving-cert\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.799113 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-trusted-ca-bundle\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.800140 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e9846351-2ac5-42ae-9332-c97459b1fa06-console-config\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.803089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e9846351-2ac5-42ae-9332-c97459b1fa06-console-oauth-config\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.803484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e9846351-2ac5-42ae-9332-c97459b1fa06-console-serving-cert\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.817353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf64q\" (UniqueName: \"kubernetes.io/projected/e9846351-2ac5-42ae-9332-c97459b1fa06-kube-api-access-kf64q\") pod \"console-6dc6c4d949-gnkpx\" (UID: \"e9846351-2ac5-42ae-9332-c97459b1fa06\") " pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.833848 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc"] Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.872460 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt"] Dec 05 11:23:07 crc kubenswrapper[4809]: W1205 11:23:07.876441 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1b67fb0_2d57_4480_9fe5_8862894ce201.slice/crio-d03e6a267d5bfb4b0c615370b04b5ee33d95e7ff159b181e5402b8b05ec8367a WatchSource:0}: Error finding container d03e6a267d5bfb4b0c615370b04b5ee33d95e7ff159b181e5402b8b05ec8367a: Status 404 returned error can't find the container with id d03e6a267d5bfb4b0c615370b04b5ee33d95e7ff159b181e5402b8b05ec8367a Dec 05 11:23:07 crc kubenswrapper[4809]: I1205 11:23:07.992042 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.103404 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/701bb0f0-123a-4b4c-b302-c43597314c22-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.111487 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/701bb0f0-123a-4b4c-b302-c43597314c22-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-cqk7j\" (UID: \"701bb0f0-123a-4b4c-b302-c43597314c22\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.282939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-8ql2q" event={"ID":"94e36ccd-7287-4326-bf5e-864dd43c7701","Type":"ContainerStarted","Data":"441a6d980ad208a778fc6b284afe8a0d245c7bcd92fd7dece67e37d9754fb912"} Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.284100 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" event={"ID":"f9d9930b-30f9-4e72-9ebe-f273ab71b519","Type":"ContainerStarted","Data":"e62c9a06154146b380a68cdbcebdf5a41b83e3cc42c181cca26aa94de12a49e3"} Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.285266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" event={"ID":"f1b67fb0-2d57-4480-9fe5-8862894ce201","Type":"ContainerStarted","Data":"d03e6a267d5bfb4b0c615370b04b5ee33d95e7ff159b181e5402b8b05ec8367a"} Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.346059 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.428065 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6dc6c4d949-gnkpx"] Dec 05 11:23:08 crc kubenswrapper[4809]: W1205 11:23:08.435588 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9846351_2ac5_42ae_9332_c97459b1fa06.slice/crio-b61427290c90ef51e470c525c37bc2d1983c94a19cea354db676c5ab58647758 WatchSource:0}: Error finding container b61427290c90ef51e470c525c37bc2d1983c94a19cea354db676c5ab58647758: Status 404 returned error can't find the container with id b61427290c90ef51e470c525c37bc2d1983c94a19cea354db676c5ab58647758 Dec 05 11:23:08 crc kubenswrapper[4809]: I1205 11:23:08.545534 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j"] Dec 05 11:23:08 crc kubenswrapper[4809]: W1205 11:23:08.551931 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod701bb0f0_123a_4b4c_b302_c43597314c22.slice/crio-89eb027c236f3a35ec6848e334a22530fe89bffdee29bb048882152c5d099bea WatchSource:0}: Error finding container 89eb027c236f3a35ec6848e334a22530fe89bffdee29bb048882152c5d099bea: Status 404 returned error can't find the container with id 89eb027c236f3a35ec6848e334a22530fe89bffdee29bb048882152c5d099bea Dec 05 11:23:09 crc kubenswrapper[4809]: I1205 11:23:09.293846 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" event={"ID":"701bb0f0-123a-4b4c-b302-c43597314c22","Type":"ContainerStarted","Data":"89eb027c236f3a35ec6848e334a22530fe89bffdee29bb048882152c5d099bea"} Dec 05 11:23:09 crc kubenswrapper[4809]: I1205 11:23:09.297992 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6dc6c4d949-gnkpx" event={"ID":"e9846351-2ac5-42ae-9332-c97459b1fa06","Type":"ContainerStarted","Data":"b6e884fe576ce51bb040f76e9e1a0d0a2849314b9cd0ec02a32d4fe80de5fb66"} Dec 05 11:23:09 crc kubenswrapper[4809]: I1205 11:23:09.298038 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6dc6c4d949-gnkpx" event={"ID":"e9846351-2ac5-42ae-9332-c97459b1fa06","Type":"ContainerStarted","Data":"b61427290c90ef51e470c525c37bc2d1983c94a19cea354db676c5ab58647758"} Dec 05 11:23:09 crc kubenswrapper[4809]: I1205 11:23:09.315847 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6dc6c4d949-gnkpx" podStartSLOduration=2.315824616 podStartE2EDuration="2.315824616s" podCreationTimestamp="2025-12-05 11:23:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:23:09.313812551 +0000 UTC m=+884.704789119" watchObservedRunningTime="2025-12-05 11:23:09.315824616 +0000 UTC m=+884.706801174" Dec 05 11:23:09 crc kubenswrapper[4809]: I1205 11:23:09.692550 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jlk9c"] Dec 05 11:23:09 crc kubenswrapper[4809]: I1205 11:23:09.693152 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jlk9c" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="registry-server" containerID="cri-o://20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297" gracePeriod=2 Dec 05 11:23:11 crc kubenswrapper[4809]: I1205 11:23:11.974627 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.062003 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44brg\" (UniqueName: \"kubernetes.io/projected/6ac37414-e8ca-4039-aa51-762842fb6bb6-kube-api-access-44brg\") pod \"6ac37414-e8ca-4039-aa51-762842fb6bb6\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.062066 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-utilities\") pod \"6ac37414-e8ca-4039-aa51-762842fb6bb6\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.062190 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-catalog-content\") pod \"6ac37414-e8ca-4039-aa51-762842fb6bb6\" (UID: \"6ac37414-e8ca-4039-aa51-762842fb6bb6\") " Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.063133 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-utilities" (OuterVolumeSpecName: "utilities") pod "6ac37414-e8ca-4039-aa51-762842fb6bb6" (UID: "6ac37414-e8ca-4039-aa51-762842fb6bb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.069739 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac37414-e8ca-4039-aa51-762842fb6bb6-kube-api-access-44brg" (OuterVolumeSpecName: "kube-api-access-44brg") pod "6ac37414-e8ca-4039-aa51-762842fb6bb6" (UID: "6ac37414-e8ca-4039-aa51-762842fb6bb6"). InnerVolumeSpecName "kube-api-access-44brg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.164446 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44brg\" (UniqueName: \"kubernetes.io/projected/6ac37414-e8ca-4039-aa51-762842fb6bb6-kube-api-access-44brg\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.164480 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.166716 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ac37414-e8ca-4039-aa51-762842fb6bb6" (UID: "6ac37414-e8ca-4039-aa51-762842fb6bb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.265397 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac37414-e8ca-4039-aa51-762842fb6bb6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.316982 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" event={"ID":"f9d9930b-30f9-4e72-9ebe-f273ab71b519","Type":"ContainerStarted","Data":"d97a18a5112508f23dda9f673119ae4046478dc7a45ad34c0268818adefb9baf"} Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.318388 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" event={"ID":"f1b67fb0-2d57-4480-9fe5-8862894ce201","Type":"ContainerStarted","Data":"c3e8f8b1e079904aaa398c2f7339de3bc66081a62be7804dc90e80ad029504e5"} Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.319179 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.321137 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-8ql2q" event={"ID":"94e36ccd-7287-4326-bf5e-864dd43c7701","Type":"ContainerStarted","Data":"6d3b7ea3947b3fa6b14cf06181472a3553459b3aab053ce2b40ce736e39bdad0"} Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.321532 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.323672 4809 generic.go:334] "Generic (PLEG): container finished" podID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerID="20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297" exitCode=0 Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.323697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlk9c" event={"ID":"6ac37414-e8ca-4039-aa51-762842fb6bb6","Type":"ContainerDied","Data":"20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297"} Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.323715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlk9c" event={"ID":"6ac37414-e8ca-4039-aa51-762842fb6bb6","Type":"ContainerDied","Data":"a9c76a519491366292ec03df7c5295719f7175da52bb057d4a645e27b110a333"} Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.323736 4809 scope.go:117] "RemoveContainer" containerID="20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.323845 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlk9c" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.338118 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" podStartSLOduration=1.236937326 podStartE2EDuration="5.338084611s" podCreationTimestamp="2025-12-05 11:23:07 +0000 UTC" firstStartedPulling="2025-12-05 11:23:07.879212994 +0000 UTC m=+883.270189572" lastFinishedPulling="2025-12-05 11:23:11.980360309 +0000 UTC m=+887.371336857" observedRunningTime="2025-12-05 11:23:12.334183565 +0000 UTC m=+887.725160133" watchObservedRunningTime="2025-12-05 11:23:12.338084611 +0000 UTC m=+887.729061169" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.352784 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-8ql2q" podStartSLOduration=1.074343197 podStartE2EDuration="5.35276701s" podCreationTimestamp="2025-12-05 11:23:07 +0000 UTC" firstStartedPulling="2025-12-05 11:23:07.700559438 +0000 UTC m=+883.091535996" lastFinishedPulling="2025-12-05 11:23:11.978983251 +0000 UTC m=+887.369959809" observedRunningTime="2025-12-05 11:23:12.350994721 +0000 UTC m=+887.741971309" watchObservedRunningTime="2025-12-05 11:23:12.35276701 +0000 UTC m=+887.743743558" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.372308 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jlk9c"] Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.377434 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jlk9c"] Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.650429 4809 scope.go:117] "RemoveContainer" containerID="4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.664143 4809 scope.go:117] "RemoveContainer" containerID="6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.734444 4809 scope.go:117] "RemoveContainer" containerID="20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297" Dec 05 11:23:12 crc kubenswrapper[4809]: E1205 11:23:12.734853 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297\": container with ID starting with 20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297 not found: ID does not exist" containerID="20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.734896 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297"} err="failed to get container status \"20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297\": rpc error: code = NotFound desc = could not find container \"20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297\": container with ID starting with 20fccf8593079584b84b615966da82cffe151781e6223a9186b2393d5ba77297 not found: ID does not exist" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.734927 4809 scope.go:117] "RemoveContainer" containerID="4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2" Dec 05 11:23:12 crc kubenswrapper[4809]: E1205 11:23:12.735259 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2\": container with ID starting with 4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2 not found: ID does not exist" containerID="4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.735322 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2"} err="failed to get container status \"4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2\": rpc error: code = NotFound desc = could not find container \"4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2\": container with ID starting with 4678a572f39092864245129198c49a845296a81c8a6e1a55b8fc0aa5ce0031f2 not found: ID does not exist" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.735368 4809 scope.go:117] "RemoveContainer" containerID="6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d" Dec 05 11:23:12 crc kubenswrapper[4809]: E1205 11:23:12.735694 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d\": container with ID starting with 6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d not found: ID does not exist" containerID="6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.735727 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d"} err="failed to get container status \"6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d\": rpc error: code = NotFound desc = could not find container \"6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d\": container with ID starting with 6ee9f1e8b391771fe8ae5ea5e667382ffe9d7544856157d854b2733a02d8b25d not found: ID does not exist" Dec 05 11:23:12 crc kubenswrapper[4809]: I1205 11:23:12.883746 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" path="/var/lib/kubelet/pods/6ac37414-e8ca-4039-aa51-762842fb6bb6/volumes" Dec 05 11:23:13 crc kubenswrapper[4809]: I1205 11:23:13.331375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" event={"ID":"701bb0f0-123a-4b4c-b302-c43597314c22","Type":"ContainerStarted","Data":"b953e4451ac9fab435f621ad6dab8a15a5bafaf5d16b1d7f2a7a79c7fa0e1367"} Dec 05 11:23:13 crc kubenswrapper[4809]: I1205 11:23:13.348254 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-cqk7j" podStartSLOduration=2.165723706 podStartE2EDuration="6.348234513s" podCreationTimestamp="2025-12-05 11:23:07 +0000 UTC" firstStartedPulling="2025-12-05 11:23:08.553914379 +0000 UTC m=+883.944890937" lastFinishedPulling="2025-12-05 11:23:12.736425186 +0000 UTC m=+888.127401744" observedRunningTime="2025-12-05 11:23:13.348190112 +0000 UTC m=+888.739166670" watchObservedRunningTime="2025-12-05 11:23:13.348234513 +0000 UTC m=+888.739211071" Dec 05 11:23:15 crc kubenswrapper[4809]: I1205 11:23:15.348872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" event={"ID":"f9d9930b-30f9-4e72-9ebe-f273ab71b519","Type":"ContainerStarted","Data":"8cd9f5b29a18962b6c41b6d6b88c96716670f54594b193efb0945363573128ab"} Dec 05 11:23:15 crc kubenswrapper[4809]: I1205 11:23:15.367193 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4wrgc" podStartSLOduration=1.560273162 podStartE2EDuration="8.367173162s" podCreationTimestamp="2025-12-05 11:23:07 +0000 UTC" firstStartedPulling="2025-12-05 11:23:07.842181317 +0000 UTC m=+883.233157875" lastFinishedPulling="2025-12-05 11:23:14.649081307 +0000 UTC m=+890.040057875" observedRunningTime="2025-12-05 11:23:15.361537699 +0000 UTC m=+890.752514257" watchObservedRunningTime="2025-12-05 11:23:15.367173162 +0000 UTC m=+890.758149720" Dec 05 11:23:17 crc kubenswrapper[4809]: I1205 11:23:17.694815 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-8ql2q" Dec 05 11:23:17 crc kubenswrapper[4809]: I1205 11:23:17.992811 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:17 crc kubenswrapper[4809]: I1205 11:23:17.992870 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:17 crc kubenswrapper[4809]: I1205 11:23:17.998577 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:18 crc kubenswrapper[4809]: I1205 11:23:18.368245 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6dc6c4d949-gnkpx" Dec 05 11:23:18 crc kubenswrapper[4809]: I1205 11:23:18.426871 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zhtzr"] Dec 05 11:23:27 crc kubenswrapper[4809]: I1205 11:23:27.650094 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pmtjt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.499133 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt"] Dec 05 11:23:40 crc kubenswrapper[4809]: E1205 11:23:40.499810 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="registry-server" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.499824 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="registry-server" Dec 05 11:23:40 crc kubenswrapper[4809]: E1205 11:23:40.499840 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="extract-content" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.499849 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="extract-content" Dec 05 11:23:40 crc kubenswrapper[4809]: E1205 11:23:40.499863 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="extract-utilities" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.499871 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="extract-utilities" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.500001 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac37414-e8ca-4039-aa51-762842fb6bb6" containerName="registry-server" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.500906 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.502698 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.520883 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt"] Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.563060 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.563218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.563294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvn7l\" (UniqueName: \"kubernetes.io/projected/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-kube-api-access-hvn7l\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.664350 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.664418 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvn7l\" (UniqueName: \"kubernetes.io/projected/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-kube-api-access-hvn7l\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.664468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.665079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.665278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.683602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvn7l\" (UniqueName: \"kubernetes.io/projected/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-kube-api-access-hvn7l\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:40 crc kubenswrapper[4809]: I1205 11:23:40.866772 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:41 crc kubenswrapper[4809]: I1205 11:23:41.239934 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt"] Dec 05 11:23:41 crc kubenswrapper[4809]: I1205 11:23:41.527217 4809 generic.go:334] "Generic (PLEG): container finished" podID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerID="48463502d4aae9a139a11ffcd088fe07ebdfab332e8d9915da541054ccab17a7" exitCode=0 Dec 05 11:23:41 crc kubenswrapper[4809]: I1205 11:23:41.527275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" event={"ID":"b272d55e-85f0-4cb6-9bf4-360b5e2efc45","Type":"ContainerDied","Data":"48463502d4aae9a139a11ffcd088fe07ebdfab332e8d9915da541054ccab17a7"} Dec 05 11:23:41 crc kubenswrapper[4809]: I1205 11:23:41.528032 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" event={"ID":"b272d55e-85f0-4cb6-9bf4-360b5e2efc45","Type":"ContainerStarted","Data":"7068ba032dd1a890319af0c57c6f00618c9bcb0ab8b94c2f14e875a9227efd83"} Dec 05 11:23:43 crc kubenswrapper[4809]: I1205 11:23:43.471205 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-zhtzr" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" containerID="cri-o://83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae" gracePeriod=15 Dec 05 11:23:43 crc kubenswrapper[4809]: I1205 11:23:43.540913 4809 generic.go:334] "Generic (PLEG): container finished" podID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerID="fdc82b1b14a801755e7913bd99b92310c66a7443c945d2f855ff44e488c71c95" exitCode=0 Dec 05 11:23:43 crc kubenswrapper[4809]: I1205 11:23:43.540975 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" event={"ID":"b272d55e-85f0-4cb6-9bf4-360b5e2efc45","Type":"ContainerDied","Data":"fdc82b1b14a801755e7913bd99b92310c66a7443c945d2f855ff44e488c71c95"} Dec 05 11:23:43 crc kubenswrapper[4809]: I1205 11:23:43.903542 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zhtzr_41d11685-f01b-4304-a08e-3c43f21e1411/console/0.log" Dec 05 11:23:43 crc kubenswrapper[4809]: I1205 11:23:43.903859 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.100911 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9fzm\" (UniqueName: \"kubernetes.io/projected/41d11685-f01b-4304-a08e-3c43f21e1411-kube-api-access-d9fzm\") pod \"41d11685-f01b-4304-a08e-3c43f21e1411\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.100987 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-trusted-ca-bundle\") pod \"41d11685-f01b-4304-a08e-3c43f21e1411\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.101020 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-oauth-config\") pod \"41d11685-f01b-4304-a08e-3c43f21e1411\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.101041 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-oauth-serving-cert\") pod \"41d11685-f01b-4304-a08e-3c43f21e1411\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.101070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-service-ca\") pod \"41d11685-f01b-4304-a08e-3c43f21e1411\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.101807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-console-config\") pod \"41d11685-f01b-4304-a08e-3c43f21e1411\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.102107 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-service-ca" (OuterVolumeSpecName: "service-ca") pod "41d11685-f01b-4304-a08e-3c43f21e1411" (UID: "41d11685-f01b-4304-a08e-3c43f21e1411"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.101928 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-console-config" (OuterVolumeSpecName: "console-config") pod "41d11685-f01b-4304-a08e-3c43f21e1411" (UID: "41d11685-f01b-4304-a08e-3c43f21e1411"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.102191 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-serving-cert\") pod \"41d11685-f01b-4304-a08e-3c43f21e1411\" (UID: \"41d11685-f01b-4304-a08e-3c43f21e1411\") " Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.102206 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "41d11685-f01b-4304-a08e-3c43f21e1411" (UID: "41d11685-f01b-4304-a08e-3c43f21e1411"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.102304 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "41d11685-f01b-4304-a08e-3c43f21e1411" (UID: "41d11685-f01b-4304-a08e-3c43f21e1411"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.103408 4809 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.103439 4809 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.103452 4809 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.103471 4809 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/41d11685-f01b-4304-a08e-3c43f21e1411-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.113837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "41d11685-f01b-4304-a08e-3c43f21e1411" (UID: "41d11685-f01b-4304-a08e-3c43f21e1411"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.116340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "41d11685-f01b-4304-a08e-3c43f21e1411" (UID: "41d11685-f01b-4304-a08e-3c43f21e1411"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.117055 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d11685-f01b-4304-a08e-3c43f21e1411-kube-api-access-d9fzm" (OuterVolumeSpecName: "kube-api-access-d9fzm") pod "41d11685-f01b-4304-a08e-3c43f21e1411" (UID: "41d11685-f01b-4304-a08e-3c43f21e1411"). InnerVolumeSpecName "kube-api-access-d9fzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.204804 4809 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.205528 4809 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/41d11685-f01b-4304-a08e-3c43f21e1411-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.205680 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9fzm\" (UniqueName: \"kubernetes.io/projected/41d11685-f01b-4304-a08e-3c43f21e1411-kube-api-access-d9fzm\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.550890 4809 generic.go:334] "Generic (PLEG): container finished" podID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerID="03d56f8707dac001b09b3b759b2b2030503344f0106c57a608ae583b16662992" exitCode=0 Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.550989 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" event={"ID":"b272d55e-85f0-4cb6-9bf4-360b5e2efc45","Type":"ContainerDied","Data":"03d56f8707dac001b09b3b759b2b2030503344f0106c57a608ae583b16662992"} Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.554560 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zhtzr_41d11685-f01b-4304-a08e-3c43f21e1411/console/0.log" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.554686 4809 generic.go:334] "Generic (PLEG): container finished" podID="41d11685-f01b-4304-a08e-3c43f21e1411" containerID="83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae" exitCode=2 Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.554734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zhtzr" event={"ID":"41d11685-f01b-4304-a08e-3c43f21e1411","Type":"ContainerDied","Data":"83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae"} Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.554788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zhtzr" event={"ID":"41d11685-f01b-4304-a08e-3c43f21e1411","Type":"ContainerDied","Data":"42d7d9ea603b91cbdb34519c68964ce30eebc14f99152efe33aaa89895004962"} Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.554818 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zhtzr" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.554827 4809 scope.go:117] "RemoveContainer" containerID="83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.581713 4809 scope.go:117] "RemoveContainer" containerID="83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae" Dec 05 11:23:44 crc kubenswrapper[4809]: E1205 11:23:44.583952 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae\": container with ID starting with 83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae not found: ID does not exist" containerID="83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.584016 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae"} err="failed to get container status \"83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae\": rpc error: code = NotFound desc = could not find container \"83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae\": container with ID starting with 83371e40df92bbfed12b0721a202a1b10a908fd7eea8c4c03d3979e53764e0ae not found: ID does not exist" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.617575 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zhtzr"] Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.623361 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-zhtzr"] Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.872184 4809 patch_prober.go:28] interesting pod/console-f9d7485db-zhtzr container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.872289 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-zhtzr" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 11:23:44 crc kubenswrapper[4809]: I1205 11:23:44.884252 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" path="/var/lib/kubelet/pods/41d11685-f01b-4304-a08e-3c43f21e1411/volumes" Dec 05 11:23:45 crc kubenswrapper[4809]: I1205 11:23:45.858595 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.027593 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-bundle\") pod \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.027670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-util\") pod \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.027698 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvn7l\" (UniqueName: \"kubernetes.io/projected/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-kube-api-access-hvn7l\") pod \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\" (UID: \"b272d55e-85f0-4cb6-9bf4-360b5e2efc45\") " Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.028454 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-bundle" (OuterVolumeSpecName: "bundle") pod "b272d55e-85f0-4cb6-9bf4-360b5e2efc45" (UID: "b272d55e-85f0-4cb6-9bf4-360b5e2efc45"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.035464 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-kube-api-access-hvn7l" (OuterVolumeSpecName: "kube-api-access-hvn7l") pod "b272d55e-85f0-4cb6-9bf4-360b5e2efc45" (UID: "b272d55e-85f0-4cb6-9bf4-360b5e2efc45"). InnerVolumeSpecName "kube-api-access-hvn7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.058887 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-util" (OuterVolumeSpecName: "util") pod "b272d55e-85f0-4cb6-9bf4-360b5e2efc45" (UID: "b272d55e-85f0-4cb6-9bf4-360b5e2efc45"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.128922 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.128971 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-util\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.128985 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvn7l\" (UniqueName: \"kubernetes.io/projected/b272d55e-85f0-4cb6-9bf4-360b5e2efc45-kube-api-access-hvn7l\") on node \"crc\" DevicePath \"\"" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.572243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" event={"ID":"b272d55e-85f0-4cb6-9bf4-360b5e2efc45","Type":"ContainerDied","Data":"7068ba032dd1a890319af0c57c6f00618c9bcb0ab8b94c2f14e875a9227efd83"} Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.572590 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7068ba032dd1a890319af0c57c6f00618c9bcb0ab8b94c2f14e875a9227efd83" Dec 05 11:23:46 crc kubenswrapper[4809]: I1205 11:23:46.572291 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.678402 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmpr"] Dec 05 11:23:53 crc kubenswrapper[4809]: E1205 11:23:53.679670 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerName="util" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.679700 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerName="util" Dec 05 11:23:53 crc kubenswrapper[4809]: E1205 11:23:53.679733 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.679753 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" Dec 05 11:23:53 crc kubenswrapper[4809]: E1205 11:23:53.679792 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerName="pull" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.679811 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerName="pull" Dec 05 11:23:53 crc kubenswrapper[4809]: E1205 11:23:53.679839 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerName="extract" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.679856 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerName="extract" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.680111 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d11685-f01b-4304-a08e-3c43f21e1411" containerName="console" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.680151 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b272d55e-85f0-4cb6-9bf4-360b5e2efc45" containerName="extract" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.681960 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.697869 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmpr"] Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.826162 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-utilities\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.826293 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k85qd\" (UniqueName: \"kubernetes.io/projected/2c1f4375-5dd6-49e8-91b7-d676bf75a728-kube-api-access-k85qd\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.826356 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-catalog-content\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.927287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k85qd\" (UniqueName: \"kubernetes.io/projected/2c1f4375-5dd6-49e8-91b7-d676bf75a728-kube-api-access-k85qd\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.927361 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-catalog-content\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.927393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-utilities\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.928226 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-catalog-content\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.928245 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-utilities\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:53 crc kubenswrapper[4809]: I1205 11:23:53.946531 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k85qd\" (UniqueName: \"kubernetes.io/projected/2c1f4375-5dd6-49e8-91b7-d676bf75a728-kube-api-access-k85qd\") pod \"redhat-marketplace-ckmpr\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:54 crc kubenswrapper[4809]: I1205 11:23:54.007958 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:23:54 crc kubenswrapper[4809]: I1205 11:23:54.276049 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmpr"] Dec 05 11:23:54 crc kubenswrapper[4809]: I1205 11:23:54.615057 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerID="8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878" exitCode=0 Dec 05 11:23:54 crc kubenswrapper[4809]: I1205 11:23:54.615186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmpr" event={"ID":"2c1f4375-5dd6-49e8-91b7-d676bf75a728","Type":"ContainerDied","Data":"8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878"} Dec 05 11:23:54 crc kubenswrapper[4809]: I1205 11:23:54.615578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmpr" event={"ID":"2c1f4375-5dd6-49e8-91b7-d676bf75a728","Type":"ContainerStarted","Data":"b740559485edd4aa45b7a39e6c0a89792596480591c08df6d7301f0b356f2e1b"} Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.895444 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-567549d75d-v268p"] Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.896256 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.901061 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.901082 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.901079 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.901650 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.901809 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-ltzns" Dec 05 11:23:55 crc kubenswrapper[4809]: I1205 11:23:55.966873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-567549d75d-v268p"] Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.052549 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-apiservice-cert\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.053071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxmt4\" (UniqueName: \"kubernetes.io/projected/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-kube-api-access-xxmt4\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.053099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-webhook-cert\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.153470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxmt4\" (UniqueName: \"kubernetes.io/projected/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-kube-api-access-xxmt4\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.153513 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-webhook-cert\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.153567 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-apiservice-cert\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.161203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-webhook-cert\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.161293 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-apiservice-cert\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.192570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxmt4\" (UniqueName: \"kubernetes.io/projected/7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9-kube-api-access-xxmt4\") pod \"metallb-operator-controller-manager-567549d75d-v268p\" (UID: \"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9\") " pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.213970 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.238379 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc"] Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.239047 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.241718 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.241877 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-v8wxx" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.241935 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.277389 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc"] Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.360330 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hspb\" (UniqueName: \"kubernetes.io/projected/e3f8c150-501c-48f7-b111-44189a425a73-kube-api-access-7hspb\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.360755 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e3f8c150-501c-48f7-b111-44189a425a73-webhook-cert\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.360786 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e3f8c150-501c-48f7-b111-44189a425a73-apiservice-cert\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.461442 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hspb\" (UniqueName: \"kubernetes.io/projected/e3f8c150-501c-48f7-b111-44189a425a73-kube-api-access-7hspb\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.461527 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e3f8c150-501c-48f7-b111-44189a425a73-webhook-cert\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.461552 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e3f8c150-501c-48f7-b111-44189a425a73-apiservice-cert\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.466491 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e3f8c150-501c-48f7-b111-44189a425a73-webhook-cert\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.467078 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e3f8c150-501c-48f7-b111-44189a425a73-apiservice-cert\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.488902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hspb\" (UniqueName: \"kubernetes.io/projected/e3f8c150-501c-48f7-b111-44189a425a73-kube-api-access-7hspb\") pod \"metallb-operator-webhook-server-86b59fcd6c-8p2pc\" (UID: \"e3f8c150-501c-48f7-b111-44189a425a73\") " pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.497348 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-567549d75d-v268p"] Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.638883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" event={"ID":"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9","Type":"ContainerStarted","Data":"a72bb092a44a12e01bba2bea90d53d1bf989bde97ad46465aed7524f61e8e117"} Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.643506 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.649759 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerID="97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80" exitCode=0 Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.649803 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmpr" event={"ID":"2c1f4375-5dd6-49e8-91b7-d676bf75a728","Type":"ContainerDied","Data":"97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80"} Dec 05 11:23:56 crc kubenswrapper[4809]: I1205 11:23:56.847953 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc"] Dec 05 11:23:56 crc kubenswrapper[4809]: W1205 11:23:56.849112 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3f8c150_501c_48f7_b111_44189a425a73.slice/crio-91d301375b48b79d43a1e9337b165ba40e340ef92a92fc2781f2841115d2673c WatchSource:0}: Error finding container 91d301375b48b79d43a1e9337b165ba40e340ef92a92fc2781f2841115d2673c: Status 404 returned error can't find the container with id 91d301375b48b79d43a1e9337b165ba40e340ef92a92fc2781f2841115d2673c Dec 05 11:23:57 crc kubenswrapper[4809]: I1205 11:23:57.658329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmpr" event={"ID":"2c1f4375-5dd6-49e8-91b7-d676bf75a728","Type":"ContainerStarted","Data":"5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2"} Dec 05 11:23:57 crc kubenswrapper[4809]: I1205 11:23:57.660478 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" event={"ID":"e3f8c150-501c-48f7-b111-44189a425a73","Type":"ContainerStarted","Data":"91d301375b48b79d43a1e9337b165ba40e340ef92a92fc2781f2841115d2673c"} Dec 05 11:23:57 crc kubenswrapper[4809]: I1205 11:23:57.681469 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ckmpr" podStartSLOduration=2.216929532 podStartE2EDuration="4.681451796s" podCreationTimestamp="2025-12-05 11:23:53 +0000 UTC" firstStartedPulling="2025-12-05 11:23:54.616460451 +0000 UTC m=+930.007437009" lastFinishedPulling="2025-12-05 11:23:57.080982715 +0000 UTC m=+932.471959273" observedRunningTime="2025-12-05 11:23:57.674213514 +0000 UTC m=+933.065190092" watchObservedRunningTime="2025-12-05 11:23:57.681451796 +0000 UTC m=+933.072428364" Dec 05 11:23:59 crc kubenswrapper[4809]: I1205 11:23:59.673736 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" event={"ID":"7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9","Type":"ContainerStarted","Data":"e907e2f8f27a4c6dd2c017d37cf9804cfabf66c8dbefd3d599b94cd437c1c96e"} Dec 05 11:23:59 crc kubenswrapper[4809]: I1205 11:23:59.674068 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:23:59 crc kubenswrapper[4809]: I1205 11:23:59.698195 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" podStartSLOduration=1.9487614469999999 podStartE2EDuration="4.698177969s" podCreationTimestamp="2025-12-05 11:23:55 +0000 UTC" firstStartedPulling="2025-12-05 11:23:56.509580385 +0000 UTC m=+931.900556943" lastFinishedPulling="2025-12-05 11:23:59.258996917 +0000 UTC m=+934.649973465" observedRunningTime="2025-12-05 11:23:59.692892979 +0000 UTC m=+935.083869557" watchObservedRunningTime="2025-12-05 11:23:59.698177969 +0000 UTC m=+935.089154527" Dec 05 11:24:03 crc kubenswrapper[4809]: I1205 11:24:03.698122 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" event={"ID":"e3f8c150-501c-48f7-b111-44189a425a73","Type":"ContainerStarted","Data":"e35c69717bede201a72b1f3f0db5319cfc926ea2f75944229b80d6d0f448afbc"} Dec 05 11:24:03 crc kubenswrapper[4809]: I1205 11:24:03.698736 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:24:03 crc kubenswrapper[4809]: I1205 11:24:03.719441 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" podStartSLOduration=1.936933284 podStartE2EDuration="7.719419814s" podCreationTimestamp="2025-12-05 11:23:56 +0000 UTC" firstStartedPulling="2025-12-05 11:23:56.852885054 +0000 UTC m=+932.243861612" lastFinishedPulling="2025-12-05 11:24:02.635371584 +0000 UTC m=+938.026348142" observedRunningTime="2025-12-05 11:24:03.716052825 +0000 UTC m=+939.107029393" watchObservedRunningTime="2025-12-05 11:24:03.719419814 +0000 UTC m=+939.110396372" Dec 05 11:24:04 crc kubenswrapper[4809]: I1205 11:24:04.008615 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:24:04 crc kubenswrapper[4809]: I1205 11:24:04.009071 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:24:04 crc kubenswrapper[4809]: I1205 11:24:04.056876 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:24:04 crc kubenswrapper[4809]: I1205 11:24:04.760100 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:24:04 crc kubenswrapper[4809]: I1205 11:24:04.856090 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmpr"] Dec 05 11:24:06 crc kubenswrapper[4809]: I1205 11:24:06.714730 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ckmpr" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="registry-server" containerID="cri-o://5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2" gracePeriod=2 Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.564047 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.614996 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k85qd\" (UniqueName: \"kubernetes.io/projected/2c1f4375-5dd6-49e8-91b7-d676bf75a728-kube-api-access-k85qd\") pod \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.615055 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-catalog-content\") pod \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.615115 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-utilities\") pod \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\" (UID: \"2c1f4375-5dd6-49e8-91b7-d676bf75a728\") " Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.616389 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-utilities" (OuterVolumeSpecName: "utilities") pod "2c1f4375-5dd6-49e8-91b7-d676bf75a728" (UID: "2c1f4375-5dd6-49e8-91b7-d676bf75a728"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.620882 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c1f4375-5dd6-49e8-91b7-d676bf75a728-kube-api-access-k85qd" (OuterVolumeSpecName: "kube-api-access-k85qd") pod "2c1f4375-5dd6-49e8-91b7-d676bf75a728" (UID: "2c1f4375-5dd6-49e8-91b7-d676bf75a728"). InnerVolumeSpecName "kube-api-access-k85qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.640884 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c1f4375-5dd6-49e8-91b7-d676bf75a728" (UID: "2c1f4375-5dd6-49e8-91b7-d676bf75a728"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.716624 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.717094 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k85qd\" (UniqueName: \"kubernetes.io/projected/2c1f4375-5dd6-49e8-91b7-d676bf75a728-kube-api-access-k85qd\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.717112 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c1f4375-5dd6-49e8-91b7-d676bf75a728-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.722708 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerID="5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2" exitCode=0 Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.722836 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmpr" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.722921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmpr" event={"ID":"2c1f4375-5dd6-49e8-91b7-d676bf75a728","Type":"ContainerDied","Data":"5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2"} Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.722963 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmpr" event={"ID":"2c1f4375-5dd6-49e8-91b7-d676bf75a728","Type":"ContainerDied","Data":"b740559485edd4aa45b7a39e6c0a89792596480591c08df6d7301f0b356f2e1b"} Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.722985 4809 scope.go:117] "RemoveContainer" containerID="5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.742940 4809 scope.go:117] "RemoveContainer" containerID="97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.755111 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmpr"] Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.761763 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmpr"] Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.768048 4809 scope.go:117] "RemoveContainer" containerID="8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.785209 4809 scope.go:117] "RemoveContainer" containerID="5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2" Dec 05 11:24:07 crc kubenswrapper[4809]: E1205 11:24:07.785740 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2\": container with ID starting with 5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2 not found: ID does not exist" containerID="5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.785793 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2"} err="failed to get container status \"5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2\": rpc error: code = NotFound desc = could not find container \"5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2\": container with ID starting with 5d3ff3413a1bc21892eb2b4bcf4f9b7f810b41fba833df6dc73bf587a041fda2 not found: ID does not exist" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.785836 4809 scope.go:117] "RemoveContainer" containerID="97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80" Dec 05 11:24:07 crc kubenswrapper[4809]: E1205 11:24:07.786232 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80\": container with ID starting with 97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80 not found: ID does not exist" containerID="97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.786264 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80"} err="failed to get container status \"97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80\": rpc error: code = NotFound desc = could not find container \"97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80\": container with ID starting with 97236f9ae033ebe0ab088e10a3a1e81a132fd63d0abf6d1a98c15f6ed4640d80 not found: ID does not exist" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.786289 4809 scope.go:117] "RemoveContainer" containerID="8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878" Dec 05 11:24:07 crc kubenswrapper[4809]: E1205 11:24:07.786582 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878\": container with ID starting with 8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878 not found: ID does not exist" containerID="8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878" Dec 05 11:24:07 crc kubenswrapper[4809]: I1205 11:24:07.786610 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878"} err="failed to get container status \"8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878\": rpc error: code = NotFound desc = could not find container \"8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878\": container with ID starting with 8a49a65e9132cbf2de5180c4fcf9faf9ad3f25164790b815344a6e26da716878 not found: ID does not exist" Dec 05 11:24:08 crc kubenswrapper[4809]: I1205 11:24:08.879827 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" path="/var/lib/kubelet/pods/2c1f4375-5dd6-49e8-91b7-d676bf75a728/volumes" Dec 05 11:24:14 crc kubenswrapper[4809]: I1205 11:24:14.046284 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:24:14 crc kubenswrapper[4809]: I1205 11:24:14.046744 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.651965 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-86b59fcd6c-8p2pc" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.961399 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8pb7q"] Dec 05 11:24:16 crc kubenswrapper[4809]: E1205 11:24:16.962311 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="registry-server" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.962370 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="registry-server" Dec 05 11:24:16 crc kubenswrapper[4809]: E1205 11:24:16.962390 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="extract-utilities" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.962399 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="extract-utilities" Dec 05 11:24:16 crc kubenswrapper[4809]: E1205 11:24:16.962454 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="extract-content" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.962467 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="extract-content" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.962781 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c1f4375-5dd6-49e8-91b7-d676bf75a728" containerName="registry-server" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.964236 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:16 crc kubenswrapper[4809]: I1205 11:24:16.975590 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8pb7q"] Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.134586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-catalog-content\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.134683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-utilities\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.134722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zspg\" (UniqueName: \"kubernetes.io/projected/e0208545-85ec-4bbf-8db5-057fc6aedc70-kube-api-access-5zspg\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.235985 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zspg\" (UniqueName: \"kubernetes.io/projected/e0208545-85ec-4bbf-8db5-057fc6aedc70-kube-api-access-5zspg\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.236113 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-catalog-content\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.236197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-utilities\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.236876 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-utilities\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.237131 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-catalog-content\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.262920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zspg\" (UniqueName: \"kubernetes.io/projected/e0208545-85ec-4bbf-8db5-057fc6aedc70-kube-api-access-5zspg\") pod \"community-operators-8pb7q\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.292380 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:17 crc kubenswrapper[4809]: I1205 11:24:17.929933 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8pb7q"] Dec 05 11:24:17 crc kubenswrapper[4809]: W1205 11:24:17.942920 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0208545_85ec_4bbf_8db5_057fc6aedc70.slice/crio-97857b8ed6fb5e1362cf3b8f3d61bc4a8d0c4efc10be0d1c74ceafc554479cf0 WatchSource:0}: Error finding container 97857b8ed6fb5e1362cf3b8f3d61bc4a8d0c4efc10be0d1c74ceafc554479cf0: Status 404 returned error can't find the container with id 97857b8ed6fb5e1362cf3b8f3d61bc4a8d0c4efc10be0d1c74ceafc554479cf0 Dec 05 11:24:18 crc kubenswrapper[4809]: I1205 11:24:18.788958 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pb7q" event={"ID":"e0208545-85ec-4bbf-8db5-057fc6aedc70","Type":"ContainerStarted","Data":"97857b8ed6fb5e1362cf3b8f3d61bc4a8d0c4efc10be0d1c74ceafc554479cf0"} Dec 05 11:24:19 crc kubenswrapper[4809]: I1205 11:24:19.795612 4809 generic.go:334] "Generic (PLEG): container finished" podID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerID="be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972" exitCode=0 Dec 05 11:24:19 crc kubenswrapper[4809]: I1205 11:24:19.795665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pb7q" event={"ID":"e0208545-85ec-4bbf-8db5-057fc6aedc70","Type":"ContainerDied","Data":"be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972"} Dec 05 11:24:21 crc kubenswrapper[4809]: I1205 11:24:21.806715 4809 generic.go:334] "Generic (PLEG): container finished" podID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerID="6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2" exitCode=0 Dec 05 11:24:21 crc kubenswrapper[4809]: I1205 11:24:21.806820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pb7q" event={"ID":"e0208545-85ec-4bbf-8db5-057fc6aedc70","Type":"ContainerDied","Data":"6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2"} Dec 05 11:24:22 crc kubenswrapper[4809]: I1205 11:24:22.831087 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pb7q" event={"ID":"e0208545-85ec-4bbf-8db5-057fc6aedc70","Type":"ContainerStarted","Data":"bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a"} Dec 05 11:24:22 crc kubenswrapper[4809]: I1205 11:24:22.849667 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8pb7q" podStartSLOduration=4.429526329 podStartE2EDuration="6.849648576s" podCreationTimestamp="2025-12-05 11:24:16 +0000 UTC" firstStartedPulling="2025-12-05 11:24:19.796782862 +0000 UTC m=+955.187759410" lastFinishedPulling="2025-12-05 11:24:22.216905099 +0000 UTC m=+957.607881657" observedRunningTime="2025-12-05 11:24:22.848994179 +0000 UTC m=+958.239970827" watchObservedRunningTime="2025-12-05 11:24:22.849648576 +0000 UTC m=+958.240625144" Dec 05 11:24:27 crc kubenswrapper[4809]: I1205 11:24:27.292881 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:27 crc kubenswrapper[4809]: I1205 11:24:27.293700 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:27 crc kubenswrapper[4809]: I1205 11:24:27.363618 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:27 crc kubenswrapper[4809]: I1205 11:24:27.898222 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:27 crc kubenswrapper[4809]: I1205 11:24:27.938839 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8pb7q"] Dec 05 11:24:29 crc kubenswrapper[4809]: I1205 11:24:29.872749 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8pb7q" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="registry-server" containerID="cri-o://bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a" gracePeriod=2 Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.244012 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.298732 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-catalog-content\") pod \"e0208545-85ec-4bbf-8db5-057fc6aedc70\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.298791 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zspg\" (UniqueName: \"kubernetes.io/projected/e0208545-85ec-4bbf-8db5-057fc6aedc70-kube-api-access-5zspg\") pod \"e0208545-85ec-4bbf-8db5-057fc6aedc70\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.299927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-utilities\") pod \"e0208545-85ec-4bbf-8db5-057fc6aedc70\" (UID: \"e0208545-85ec-4bbf-8db5-057fc6aedc70\") " Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.301089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-utilities" (OuterVolumeSpecName: "utilities") pod "e0208545-85ec-4bbf-8db5-057fc6aedc70" (UID: "e0208545-85ec-4bbf-8db5-057fc6aedc70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.305542 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0208545-85ec-4bbf-8db5-057fc6aedc70-kube-api-access-5zspg" (OuterVolumeSpecName: "kube-api-access-5zspg") pod "e0208545-85ec-4bbf-8db5-057fc6aedc70" (UID: "e0208545-85ec-4bbf-8db5-057fc6aedc70"). InnerVolumeSpecName "kube-api-access-5zspg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.351062 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0208545-85ec-4bbf-8db5-057fc6aedc70" (UID: "e0208545-85ec-4bbf-8db5-057fc6aedc70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.401775 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.402075 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zspg\" (UniqueName: \"kubernetes.io/projected/e0208545-85ec-4bbf-8db5-057fc6aedc70-kube-api-access-5zspg\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.402134 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0208545-85ec-4bbf-8db5-057fc6aedc70-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.889083 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pb7q" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.890992 4809 generic.go:334] "Generic (PLEG): container finished" podID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerID="bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a" exitCode=0 Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.891048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pb7q" event={"ID":"e0208545-85ec-4bbf-8db5-057fc6aedc70","Type":"ContainerDied","Data":"bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a"} Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.891104 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pb7q" event={"ID":"e0208545-85ec-4bbf-8db5-057fc6aedc70","Type":"ContainerDied","Data":"97857b8ed6fb5e1362cf3b8f3d61bc4a8d0c4efc10be0d1c74ceafc554479cf0"} Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.891126 4809 scope.go:117] "RemoveContainer" containerID="bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.932210 4809 scope.go:117] "RemoveContainer" containerID="6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.979068 4809 scope.go:117] "RemoveContainer" containerID="be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972" Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.981103 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8pb7q"] Dec 05 11:24:30 crc kubenswrapper[4809]: I1205 11:24:30.994697 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8pb7q"] Dec 05 11:24:31 crc kubenswrapper[4809]: I1205 11:24:31.001492 4809 scope.go:117] "RemoveContainer" containerID="bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a" Dec 05 11:24:31 crc kubenswrapper[4809]: E1205 11:24:31.002213 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a\": container with ID starting with bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a not found: ID does not exist" containerID="bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a" Dec 05 11:24:31 crc kubenswrapper[4809]: I1205 11:24:31.002323 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a"} err="failed to get container status \"bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a\": rpc error: code = NotFound desc = could not find container \"bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a\": container with ID starting with bcd62ccc01b117cf9966329c462333c80af86a824a4ab08b3be53831e41b4c4a not found: ID does not exist" Dec 05 11:24:31 crc kubenswrapper[4809]: I1205 11:24:31.002395 4809 scope.go:117] "RemoveContainer" containerID="6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2" Dec 05 11:24:31 crc kubenswrapper[4809]: E1205 11:24:31.003178 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2\": container with ID starting with 6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2 not found: ID does not exist" containerID="6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2" Dec 05 11:24:31 crc kubenswrapper[4809]: I1205 11:24:31.003217 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2"} err="failed to get container status \"6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2\": rpc error: code = NotFound desc = could not find container \"6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2\": container with ID starting with 6e6d85f235453a2578969202daf422b31dff41097b05c1f6fb299371311fa1f2 not found: ID does not exist" Dec 05 11:24:31 crc kubenswrapper[4809]: I1205 11:24:31.003275 4809 scope.go:117] "RemoveContainer" containerID="be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972" Dec 05 11:24:31 crc kubenswrapper[4809]: E1205 11:24:31.003591 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972\": container with ID starting with be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972 not found: ID does not exist" containerID="be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972" Dec 05 11:24:31 crc kubenswrapper[4809]: I1205 11:24:31.003760 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972"} err="failed to get container status \"be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972\": rpc error: code = NotFound desc = could not find container \"be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972\": container with ID starting with be49cdf510c52b48cf186adcbc63bc91ea7f6ee0da94dcae0df735c985152972 not found: ID does not exist" Dec 05 11:24:32 crc kubenswrapper[4809]: I1205 11:24:32.885130 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" path="/var/lib/kubelet/pods/e0208545-85ec-4bbf-8db5-057fc6aedc70/volumes" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.216381 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-567549d75d-v268p" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.960275 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-89669"] Dec 05 11:24:36 crc kubenswrapper[4809]: E1205 11:24:36.960735 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="extract-utilities" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.960751 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="extract-utilities" Dec 05 11:24:36 crc kubenswrapper[4809]: E1205 11:24:36.960765 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="registry-server" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.960771 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="registry-server" Dec 05 11:24:36 crc kubenswrapper[4809]: E1205 11:24:36.960783 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="extract-content" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.960789 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="extract-content" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.960879 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0208545-85ec-4bbf-8db5-057fc6aedc70" containerName="registry-server" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.961269 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.965196 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.965586 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dtpgs" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.965929 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-pgs44"] Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.969301 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.974333 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.974559 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 05 11:24:36 crc kubenswrapper[4809]: I1205 11:24:36.978812 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-89669"] Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.051974 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-ns7rn"] Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.053257 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.055291 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.055455 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.055604 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-sqp94" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.055976 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.057184 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-hqh9p"] Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.058268 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.059357 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.076105 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-hqh9p"] Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119741 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm4hs\" (UniqueName: \"kubernetes.io/projected/7f08f73d-94ad-41ad-8406-db20168bb7cd-kube-api-access-mm4hs\") pod \"frr-k8s-webhook-server-7fcb986d4-89669\" (UID: \"7f08f73d-94ad-41ad-8406-db20168bb7cd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119809 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-conf\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-sockets\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119881 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p57r\" (UniqueName: \"kubernetes.io/projected/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-kube-api-access-4p57r\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119912 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f08f73d-94ad-41ad-8406-db20168bb7cd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-89669\" (UID: \"7f08f73d-94ad-41ad-8406-db20168bb7cd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-metrics-certs\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-reloader\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.119989 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-metrics\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.120008 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-startup\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4mfr\" (UniqueName: \"kubernetes.io/projected/aaa44bcc-a21a-45db-a639-c8336476e155-kube-api-access-f4mfr\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221433 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-metrics\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-startup\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221529 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm4hs\" (UniqueName: \"kubernetes.io/projected/7f08f73d-94ad-41ad-8406-db20168bb7cd-kube-api-access-mm4hs\") pod \"frr-k8s-webhook-server-7fcb986d4-89669\" (UID: \"7f08f73d-94ad-41ad-8406-db20168bb7cd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221560 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aaa44bcc-a21a-45db-a639-c8336476e155-metrics-certs\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-conf\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa44bcc-a21a-45db-a639-c8336476e155-cert\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-sockets\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-metrics-certs\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-memberlist\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221726 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p57r\" (UniqueName: \"kubernetes.io/projected/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-kube-api-access-4p57r\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f08f73d-94ad-41ad-8406-db20168bb7cd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-89669\" (UID: \"7f08f73d-94ad-41ad-8406-db20168bb7cd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221782 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-metrics-certs\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221811 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xszmc\" (UniqueName: \"kubernetes.io/projected/57d73721-c422-46ba-b347-b28c8027351a-kube-api-access-xszmc\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-reloader\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.221848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/57d73721-c422-46ba-b347-b28c8027351a-metallb-excludel2\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.222053 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-metrics\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.222102 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-conf\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.222536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-sockets\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.222904 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-reloader\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.223101 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-frr-startup\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.232221 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-metrics-certs\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.232279 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f08f73d-94ad-41ad-8406-db20168bb7cd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-89669\" (UID: \"7f08f73d-94ad-41ad-8406-db20168bb7cd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.238894 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p57r\" (UniqueName: \"kubernetes.io/projected/864e3e5e-8812-4482-8cf0-a36c2e3c4d36-kube-api-access-4p57r\") pod \"frr-k8s-pgs44\" (UID: \"864e3e5e-8812-4482-8cf0-a36c2e3c4d36\") " pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.256253 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm4hs\" (UniqueName: \"kubernetes.io/projected/7f08f73d-94ad-41ad-8406-db20168bb7cd-kube-api-access-mm4hs\") pod \"frr-k8s-webhook-server-7fcb986d4-89669\" (UID: \"7f08f73d-94ad-41ad-8406-db20168bb7cd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.277171 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.294357 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.327731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-metrics-certs\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.328051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-memberlist\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.328099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xszmc\" (UniqueName: \"kubernetes.io/projected/57d73721-c422-46ba-b347-b28c8027351a-kube-api-access-xszmc\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.328119 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/57d73721-c422-46ba-b347-b28c8027351a-metallb-excludel2\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.328138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4mfr\" (UniqueName: \"kubernetes.io/projected/aaa44bcc-a21a-45db-a639-c8336476e155-kube-api-access-f4mfr\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.328171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aaa44bcc-a21a-45db-a639-c8336476e155-metrics-certs\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.328197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa44bcc-a21a-45db-a639-c8336476e155-cert\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: E1205 11:24:37.328317 4809 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 11:24:37 crc kubenswrapper[4809]: E1205 11:24:37.328366 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-memberlist podName:57d73721-c422-46ba-b347-b28c8027351a nodeName:}" failed. No retries permitted until 2025-12-05 11:24:37.82834962 +0000 UTC m=+973.219326178 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-memberlist") pod "speaker-ns7rn" (UID: "57d73721-c422-46ba-b347-b28c8027351a") : secret "metallb-memberlist" not found Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.329261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/57d73721-c422-46ba-b347-b28c8027351a-metallb-excludel2\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.331614 4809 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.331866 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aaa44bcc-a21a-45db-a639-c8336476e155-metrics-certs\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.333007 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-metrics-certs\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.342845 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa44bcc-a21a-45db-a639-c8336476e155-cert\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.346969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4mfr\" (UniqueName: \"kubernetes.io/projected/aaa44bcc-a21a-45db-a639-c8336476e155-kube-api-access-f4mfr\") pod \"controller-f8648f98b-hqh9p\" (UID: \"aaa44bcc-a21a-45db-a639-c8336476e155\") " pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.352142 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xszmc\" (UniqueName: \"kubernetes.io/projected/57d73721-c422-46ba-b347-b28c8027351a-kube-api-access-xszmc\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.418078 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.498447 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-89669"] Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.614694 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-hqh9p"] Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.835941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-memberlist\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.845190 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/57d73721-c422-46ba-b347-b28c8027351a-memberlist\") pod \"speaker-ns7rn\" (UID: \"57d73721-c422-46ba-b347-b28c8027351a\") " pod="metallb-system/speaker-ns7rn" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.935982 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" event={"ID":"7f08f73d-94ad-41ad-8406-db20168bb7cd","Type":"ContainerStarted","Data":"4f151765bd074d6898dc8aabb7b7ef5b0731b4729b10dd4f18e7b05099d04f56"} Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.938088 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hqh9p" event={"ID":"aaa44bcc-a21a-45db-a639-c8336476e155","Type":"ContainerStarted","Data":"a5ad12787f5b35d494de173829965d9d7641d5e64b4c3cb721ef2cd139a44278"} Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.938123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hqh9p" event={"ID":"aaa44bcc-a21a-45db-a639-c8336476e155","Type":"ContainerStarted","Data":"bb7af8c7e3c43df6a43874e3d7b395de90a0575a476e1013f8596b7e609f917e"} Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.938166 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hqh9p" event={"ID":"aaa44bcc-a21a-45db-a639-c8336476e155","Type":"ContainerStarted","Data":"ff758012e0bcd286e023c1ffb0ed8f545276f54e72a19876c9986de84725f5ed"} Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.938287 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.939416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerStarted","Data":"c51f05de0bdcd4d95095bff6332b4c1dc4409ea6fcb4555c12ea9d8e18c018ea"} Dec 05 11:24:37 crc kubenswrapper[4809]: I1205 11:24:37.959048 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-hqh9p" podStartSLOduration=0.959033382 podStartE2EDuration="959.033382ms" podCreationTimestamp="2025-12-05 11:24:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:24:37.957766118 +0000 UTC m=+973.348742676" watchObservedRunningTime="2025-12-05 11:24:37.959033382 +0000 UTC m=+973.350009940" Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.000398 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ns7rn" Dec 05 11:24:38 crc kubenswrapper[4809]: W1205 11:24:38.026987 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57d73721_c422_46ba_b347_b28c8027351a.slice/crio-2dbbe35254e05e69460c7e8aaa4a0f413c75b7d20f394bba120b4890540d8d04 WatchSource:0}: Error finding container 2dbbe35254e05e69460c7e8aaa4a0f413c75b7d20f394bba120b4890540d8d04: Status 404 returned error can't find the container with id 2dbbe35254e05e69460c7e8aaa4a0f413c75b7d20f394bba120b4890540d8d04 Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.937816 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vpj47"] Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.939815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.955199 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ns7rn" event={"ID":"57d73721-c422-46ba-b347-b28c8027351a","Type":"ContainerStarted","Data":"d5c803ea319b8edb126e358e91320f0ec230d72679828a4fee624e3f227f4ba4"} Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.955255 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ns7rn" event={"ID":"57d73721-c422-46ba-b347-b28c8027351a","Type":"ContainerStarted","Data":"ab6278da9cd8606ca0f36c44727af0da71f01cea427944a3bdba9dd504ac1a11"} Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.955268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ns7rn" event={"ID":"57d73721-c422-46ba-b347-b28c8027351a","Type":"ContainerStarted","Data":"2dbbe35254e05e69460c7e8aaa4a0f413c75b7d20f394bba120b4890540d8d04"} Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.955499 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-ns7rn" Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.963060 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vpj47"] Dec 05 11:24:38 crc kubenswrapper[4809]: I1205 11:24:38.995006 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-ns7rn" podStartSLOduration=1.994987666 podStartE2EDuration="1.994987666s" podCreationTimestamp="2025-12-05 11:24:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:24:38.991689289 +0000 UTC m=+974.382665847" watchObservedRunningTime="2025-12-05 11:24:38.994987666 +0000 UTC m=+974.385964224" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.066319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-catalog-content\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.066597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n729\" (UniqueName: \"kubernetes.io/projected/e612454b-3e90-481c-aafa-814057d56371-kube-api-access-9n729\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.066749 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-utilities\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.167219 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n729\" (UniqueName: \"kubernetes.io/projected/e612454b-3e90-481c-aafa-814057d56371-kube-api-access-9n729\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.167262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-catalog-content\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.167300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-utilities\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.167747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-utilities\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.167891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-catalog-content\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.189776 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n729\" (UniqueName: \"kubernetes.io/projected/e612454b-3e90-481c-aafa-814057d56371-kube-api-access-9n729\") pod \"certified-operators-vpj47\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.261899 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.534320 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vpj47"] Dec 05 11:24:39 crc kubenswrapper[4809]: W1205 11:24:39.553931 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode612454b_3e90_481c_aafa_814057d56371.slice/crio-4231d3bf86c8a0e9d8d9d585c4b5944c247510d1cb18e776408c3da622903184 WatchSource:0}: Error finding container 4231d3bf86c8a0e9d8d9d585c4b5944c247510d1cb18e776408c3da622903184: Status 404 returned error can't find the container with id 4231d3bf86c8a0e9d8d9d585c4b5944c247510d1cb18e776408c3da622903184 Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.963190 4809 generic.go:334] "Generic (PLEG): container finished" podID="e612454b-3e90-481c-aafa-814057d56371" containerID="ec032382d2086b57f35a1dca710b33ef6303bbb7ff7a6c7b3b49ea186eef565c" exitCode=0 Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.963302 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vpj47" event={"ID":"e612454b-3e90-481c-aafa-814057d56371","Type":"ContainerDied","Data":"ec032382d2086b57f35a1dca710b33ef6303bbb7ff7a6c7b3b49ea186eef565c"} Dec 05 11:24:39 crc kubenswrapper[4809]: I1205 11:24:39.963355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vpj47" event={"ID":"e612454b-3e90-481c-aafa-814057d56371","Type":"ContainerStarted","Data":"4231d3bf86c8a0e9d8d9d585c4b5944c247510d1cb18e776408c3da622903184"} Dec 05 11:24:40 crc kubenswrapper[4809]: I1205 11:24:40.971580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vpj47" event={"ID":"e612454b-3e90-481c-aafa-814057d56371","Type":"ContainerStarted","Data":"da152d2e6b04115e2dd642994cfd574f33f9f2aa25841aed95908cb013743c14"} Dec 05 11:24:41 crc kubenswrapper[4809]: I1205 11:24:41.978367 4809 generic.go:334] "Generic (PLEG): container finished" podID="e612454b-3e90-481c-aafa-814057d56371" containerID="da152d2e6b04115e2dd642994cfd574f33f9f2aa25841aed95908cb013743c14" exitCode=0 Dec 05 11:24:41 crc kubenswrapper[4809]: I1205 11:24:41.978411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vpj47" event={"ID":"e612454b-3e90-481c-aafa-814057d56371","Type":"ContainerDied","Data":"da152d2e6b04115e2dd642994cfd574f33f9f2aa25841aed95908cb013743c14"} Dec 05 11:24:44 crc kubenswrapper[4809]: I1205 11:24:44.047544 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:24:44 crc kubenswrapper[4809]: I1205 11:24:44.047681 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:24:45 crc kubenswrapper[4809]: I1205 11:24:44.999798 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" event={"ID":"7f08f73d-94ad-41ad-8406-db20168bb7cd","Type":"ContainerStarted","Data":"08c273d08b1059d95ef51bf13f7ef4592f03e97c7ac3f3130ffff510cde7877f"} Dec 05 11:24:45 crc kubenswrapper[4809]: I1205 11:24:45.001121 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:45 crc kubenswrapper[4809]: I1205 11:24:45.003542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vpj47" event={"ID":"e612454b-3e90-481c-aafa-814057d56371","Type":"ContainerStarted","Data":"c5fede97bc2c5ba4e57ea14dcdb5074d07f212c0f31a135f48fdf74a4e32372a"} Dec 05 11:24:45 crc kubenswrapper[4809]: I1205 11:24:45.020172 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" podStartSLOduration=2.316986389 podStartE2EDuration="9.020152286s" podCreationTimestamp="2025-12-05 11:24:36 +0000 UTC" firstStartedPulling="2025-12-05 11:24:37.506016534 +0000 UTC m=+972.896993092" lastFinishedPulling="2025-12-05 11:24:44.209182431 +0000 UTC m=+979.600158989" observedRunningTime="2025-12-05 11:24:45.016245122 +0000 UTC m=+980.407221690" watchObservedRunningTime="2025-12-05 11:24:45.020152286 +0000 UTC m=+980.411128844" Dec 05 11:24:45 crc kubenswrapper[4809]: I1205 11:24:45.039078 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vpj47" podStartSLOduration=2.478241526 podStartE2EDuration="7.039063397s" podCreationTimestamp="2025-12-05 11:24:38 +0000 UTC" firstStartedPulling="2025-12-05 11:24:39.964855027 +0000 UTC m=+975.355831575" lastFinishedPulling="2025-12-05 11:24:44.525676878 +0000 UTC m=+979.916653446" observedRunningTime="2025-12-05 11:24:45.036941191 +0000 UTC m=+980.427917749" watchObservedRunningTime="2025-12-05 11:24:45.039063397 +0000 UTC m=+980.430039945" Dec 05 11:24:47 crc kubenswrapper[4809]: I1205 11:24:47.017069 4809 generic.go:334] "Generic (PLEG): container finished" podID="864e3e5e-8812-4482-8cf0-a36c2e3c4d36" containerID="b9dcf074b2dae3f9edaff0f67f3937091e81a62bce8a08fbf769d8facd978a4b" exitCode=0 Dec 05 11:24:47 crc kubenswrapper[4809]: I1205 11:24:47.017160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerDied","Data":"b9dcf074b2dae3f9edaff0f67f3937091e81a62bce8a08fbf769d8facd978a4b"} Dec 05 11:24:47 crc kubenswrapper[4809]: I1205 11:24:47.424548 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-hqh9p" Dec 05 11:24:48 crc kubenswrapper[4809]: I1205 11:24:48.004396 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-ns7rn" Dec 05 11:24:48 crc kubenswrapper[4809]: I1205 11:24:48.027441 4809 generic.go:334] "Generic (PLEG): container finished" podID="864e3e5e-8812-4482-8cf0-a36c2e3c4d36" containerID="1d1f80fb74a25eb3e2543b5ac9ef10848b9b09d206801fd69cd563a2da63669d" exitCode=0 Dec 05 11:24:48 crc kubenswrapper[4809]: I1205 11:24:48.027487 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerDied","Data":"1d1f80fb74a25eb3e2543b5ac9ef10848b9b09d206801fd69cd563a2da63669d"} Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.038321 4809 generic.go:334] "Generic (PLEG): container finished" podID="864e3e5e-8812-4482-8cf0-a36c2e3c4d36" containerID="f78f2e43b4ce4f3a7b87bcfbf682be574d1c13aeb489a436654a4bd9876fba0e" exitCode=0 Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.038387 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerDied","Data":"f78f2e43b4ce4f3a7b87bcfbf682be574d1c13aeb489a436654a4bd9876fba0e"} Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.262209 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.262257 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.298249 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.583901 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j"] Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.584947 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.593133 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.630543 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j"] Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.655397 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x2wl\" (UniqueName: \"kubernetes.io/projected/2455a23f-14d3-41b8-9bef-e75ee783186a-kube-api-access-6x2wl\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.655460 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.655531 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.756860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x2wl\" (UniqueName: \"kubernetes.io/projected/2455a23f-14d3-41b8-9bef-e75ee783186a-kube-api-access-6x2wl\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.756920 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.756979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.757436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.757667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.780151 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x2wl\" (UniqueName: \"kubernetes.io/projected/2455a23f-14d3-41b8-9bef-e75ee783186a-kube-api-access-6x2wl\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:49 crc kubenswrapper[4809]: I1205 11:24:49.898585 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:50 crc kubenswrapper[4809]: I1205 11:24:50.052514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerStarted","Data":"c959f20be86e27d900adde2fdf78fa0d35cc61710aaeadbde7fb7fa9f41eea19"} Dec 05 11:24:50 crc kubenswrapper[4809]: I1205 11:24:50.052823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerStarted","Data":"2e882d0a23abbd3a6ed4544e275757417e2d97b5254bd2bce75bcb2fe1f415e3"} Dec 05 11:24:50 crc kubenswrapper[4809]: I1205 11:24:50.052841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerStarted","Data":"49c30609761f132e9327f9dcd03ab72f4243f78cd9e3adaf6d176082400d539b"} Dec 05 11:24:50 crc kubenswrapper[4809]: I1205 11:24:50.052853 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerStarted","Data":"37255fd0fdb682de5901f2f32a916c9367ca79b34785cbfc71b196127fead3ac"} Dec 05 11:24:50 crc kubenswrapper[4809]: I1205 11:24:50.052864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerStarted","Data":"8e8182a0caef378191865f048edea8f15db239dc90f069dde4431640e1f4cce1"} Dec 05 11:24:50 crc kubenswrapper[4809]: I1205 11:24:50.109927 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:50 crc kubenswrapper[4809]: I1205 11:24:50.138233 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j"] Dec 05 11:24:51 crc kubenswrapper[4809]: I1205 11:24:51.064348 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgs44" event={"ID":"864e3e5e-8812-4482-8cf0-a36c2e3c4d36","Type":"ContainerStarted","Data":"d2121759e8e5f839cfab200d2fe54af1938a3fd3f81f578c034fb890af5b4b2a"} Dec 05 11:24:51 crc kubenswrapper[4809]: I1205 11:24:51.064681 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:51 crc kubenswrapper[4809]: I1205 11:24:51.066171 4809 generic.go:334] "Generic (PLEG): container finished" podID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerID="fb9be3427a8f35dd380789482013150b6743b2d4de56221232a611ef16a4c9a5" exitCode=0 Dec 05 11:24:51 crc kubenswrapper[4809]: I1205 11:24:51.066266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" event={"ID":"2455a23f-14d3-41b8-9bef-e75ee783186a","Type":"ContainerDied","Data":"fb9be3427a8f35dd380789482013150b6743b2d4de56221232a611ef16a4c9a5"} Dec 05 11:24:51 crc kubenswrapper[4809]: I1205 11:24:51.066322 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" event={"ID":"2455a23f-14d3-41b8-9bef-e75ee783186a","Type":"ContainerStarted","Data":"aafd40528c3804205be5248741b35b76b4d47a36c1b0d2ab7819f221a7bc6ffd"} Dec 05 11:24:51 crc kubenswrapper[4809]: I1205 11:24:51.094320 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-pgs44" podStartSLOduration=6.079634454 podStartE2EDuration="15.094301846s" podCreationTimestamp="2025-12-05 11:24:36 +0000 UTC" firstStartedPulling="2025-12-05 11:24:37.497800326 +0000 UTC m=+972.888776884" lastFinishedPulling="2025-12-05 11:24:46.512467718 +0000 UTC m=+981.903444276" observedRunningTime="2025-12-05 11:24:51.092771675 +0000 UTC m=+986.483748243" watchObservedRunningTime="2025-12-05 11:24:51.094301846 +0000 UTC m=+986.485278404" Dec 05 11:24:52 crc kubenswrapper[4809]: I1205 11:24:52.295164 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:52 crc kubenswrapper[4809]: I1205 11:24:52.390253 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-pgs44" Dec 05 11:24:52 crc kubenswrapper[4809]: I1205 11:24:52.742060 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vpj47"] Dec 05 11:24:52 crc kubenswrapper[4809]: I1205 11:24:52.742361 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vpj47" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="registry-server" containerID="cri-o://c5fede97bc2c5ba4e57ea14dcdb5074d07f212c0f31a135f48fdf74a4e32372a" gracePeriod=2 Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.080366 4809 generic.go:334] "Generic (PLEG): container finished" podID="e612454b-3e90-481c-aafa-814057d56371" containerID="c5fede97bc2c5ba4e57ea14dcdb5074d07f212c0f31a135f48fdf74a4e32372a" exitCode=0 Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.080472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vpj47" event={"ID":"e612454b-3e90-481c-aafa-814057d56371","Type":"ContainerDied","Data":"c5fede97bc2c5ba4e57ea14dcdb5074d07f212c0f31a135f48fdf74a4e32372a"} Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.354325 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.405677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n729\" (UniqueName: \"kubernetes.io/projected/e612454b-3e90-481c-aafa-814057d56371-kube-api-access-9n729\") pod \"e612454b-3e90-481c-aafa-814057d56371\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.405778 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-utilities\") pod \"e612454b-3e90-481c-aafa-814057d56371\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.405873 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-catalog-content\") pod \"e612454b-3e90-481c-aafa-814057d56371\" (UID: \"e612454b-3e90-481c-aafa-814057d56371\") " Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.406940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-utilities" (OuterVolumeSpecName: "utilities") pod "e612454b-3e90-481c-aafa-814057d56371" (UID: "e612454b-3e90-481c-aafa-814057d56371"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.412912 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e612454b-3e90-481c-aafa-814057d56371-kube-api-access-9n729" (OuterVolumeSpecName: "kube-api-access-9n729") pod "e612454b-3e90-481c-aafa-814057d56371" (UID: "e612454b-3e90-481c-aafa-814057d56371"). InnerVolumeSpecName "kube-api-access-9n729". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.460219 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e612454b-3e90-481c-aafa-814057d56371" (UID: "e612454b-3e90-481c-aafa-814057d56371"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.507580 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n729\" (UniqueName: \"kubernetes.io/projected/e612454b-3e90-481c-aafa-814057d56371-kube-api-access-9n729\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.507654 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:53 crc kubenswrapper[4809]: I1205 11:24:53.507675 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e612454b-3e90-481c-aafa-814057d56371-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:54 crc kubenswrapper[4809]: I1205 11:24:54.091580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vpj47" event={"ID":"e612454b-3e90-481c-aafa-814057d56371","Type":"ContainerDied","Data":"4231d3bf86c8a0e9d8d9d585c4b5944c247510d1cb18e776408c3da622903184"} Dec 05 11:24:54 crc kubenswrapper[4809]: I1205 11:24:54.091647 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vpj47" Dec 05 11:24:54 crc kubenswrapper[4809]: I1205 11:24:54.091689 4809 scope.go:117] "RemoveContainer" containerID="c5fede97bc2c5ba4e57ea14dcdb5074d07f212c0f31a135f48fdf74a4e32372a" Dec 05 11:24:54 crc kubenswrapper[4809]: I1205 11:24:54.122272 4809 scope.go:117] "RemoveContainer" containerID="da152d2e6b04115e2dd642994cfd574f33f9f2aa25841aed95908cb013743c14" Dec 05 11:24:54 crc kubenswrapper[4809]: I1205 11:24:54.126482 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vpj47"] Dec 05 11:24:54 crc kubenswrapper[4809]: I1205 11:24:54.131970 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vpj47"] Dec 05 11:24:54 crc kubenswrapper[4809]: I1205 11:24:54.881323 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e612454b-3e90-481c-aafa-814057d56371" path="/var/lib/kubelet/pods/e612454b-3e90-481c-aafa-814057d56371/volumes" Dec 05 11:24:55 crc kubenswrapper[4809]: I1205 11:24:55.279054 4809 scope.go:117] "RemoveContainer" containerID="ec032382d2086b57f35a1dca710b33ef6303bbb7ff7a6c7b3b49ea186eef565c" Dec 05 11:24:56 crc kubenswrapper[4809]: I1205 11:24:56.114957 4809 generic.go:334] "Generic (PLEG): container finished" podID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerID="332d7f0cd11b113942e81d8506bc48a5b723a343b9de3c82691cc98afe38449e" exitCode=0 Dec 05 11:24:56 crc kubenswrapper[4809]: I1205 11:24:56.115804 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" event={"ID":"2455a23f-14d3-41b8-9bef-e75ee783186a","Type":"ContainerDied","Data":"332d7f0cd11b113942e81d8506bc48a5b723a343b9de3c82691cc98afe38449e"} Dec 05 11:24:57 crc kubenswrapper[4809]: I1205 11:24:57.125094 4809 generic.go:334] "Generic (PLEG): container finished" podID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerID="c4f7556407559a1fffa674d6a3356d082966a3fedc592784823cff26121cbe64" exitCode=0 Dec 05 11:24:57 crc kubenswrapper[4809]: I1205 11:24:57.125139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" event={"ID":"2455a23f-14d3-41b8-9bef-e75ee783186a","Type":"ContainerDied","Data":"c4f7556407559a1fffa674d6a3356d082966a3fedc592784823cff26121cbe64"} Dec 05 11:24:57 crc kubenswrapper[4809]: I1205 11:24:57.283708 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-89669" Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.607440 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.678207 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-bundle\") pod \"2455a23f-14d3-41b8-9bef-e75ee783186a\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.678304 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-util\") pod \"2455a23f-14d3-41b8-9bef-e75ee783186a\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.678337 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x2wl\" (UniqueName: \"kubernetes.io/projected/2455a23f-14d3-41b8-9bef-e75ee783186a-kube-api-access-6x2wl\") pod \"2455a23f-14d3-41b8-9bef-e75ee783186a\" (UID: \"2455a23f-14d3-41b8-9bef-e75ee783186a\") " Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.679395 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-bundle" (OuterVolumeSpecName: "bundle") pod "2455a23f-14d3-41b8-9bef-e75ee783186a" (UID: "2455a23f-14d3-41b8-9bef-e75ee783186a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.683620 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2455a23f-14d3-41b8-9bef-e75ee783186a-kube-api-access-6x2wl" (OuterVolumeSpecName: "kube-api-access-6x2wl") pod "2455a23f-14d3-41b8-9bef-e75ee783186a" (UID: "2455a23f-14d3-41b8-9bef-e75ee783186a"). InnerVolumeSpecName "kube-api-access-6x2wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.689189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-util" (OuterVolumeSpecName: "util") pod "2455a23f-14d3-41b8-9bef-e75ee783186a" (UID: "2455a23f-14d3-41b8-9bef-e75ee783186a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.779964 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.779999 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2455a23f-14d3-41b8-9bef-e75ee783186a-util\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:58 crc kubenswrapper[4809]: I1205 11:24:58.780013 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x2wl\" (UniqueName: \"kubernetes.io/projected/2455a23f-14d3-41b8-9bef-e75ee783186a-kube-api-access-6x2wl\") on node \"crc\" DevicePath \"\"" Dec 05 11:24:59 crc kubenswrapper[4809]: I1205 11:24:59.143175 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" event={"ID":"2455a23f-14d3-41b8-9bef-e75ee783186a","Type":"ContainerDied","Data":"aafd40528c3804205be5248741b35b76b4d47a36c1b0d2ab7819f221a7bc6ffd"} Dec 05 11:24:59 crc kubenswrapper[4809]: I1205 11:24:59.143217 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aafd40528c3804205be5248741b35b76b4d47a36c1b0d2ab7819f221a7bc6ffd" Dec 05 11:24:59 crc kubenswrapper[4809]: I1205 11:24:59.143303 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.240357 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75"] Dec 05 11:25:03 crc kubenswrapper[4809]: E1205 11:25:03.241349 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerName="pull" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241374 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerName="pull" Dec 05 11:25:03 crc kubenswrapper[4809]: E1205 11:25:03.241399 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerName="util" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241410 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerName="util" Dec 05 11:25:03 crc kubenswrapper[4809]: E1205 11:25:03.241427 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="extract-content" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241440 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="extract-content" Dec 05 11:25:03 crc kubenswrapper[4809]: E1205 11:25:03.241463 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="registry-server" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241474 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="registry-server" Dec 05 11:25:03 crc kubenswrapper[4809]: E1205 11:25:03.241499 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerName="extract" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241511 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerName="extract" Dec 05 11:25:03 crc kubenswrapper[4809]: E1205 11:25:03.241525 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="extract-utilities" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241536 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="extract-utilities" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241765 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2455a23f-14d3-41b8-9bef-e75ee783186a" containerName="extract" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.241783 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e612454b-3e90-481c-aafa-814057d56371" containerName="registry-server" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.242460 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.245952 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.246875 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-gp9jw" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.247086 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.263443 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75"] Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.344251 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/32057f35-2096-4a60-be4d-24171efc379f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n8x75\" (UID: \"32057f35-2096-4a60-be4d-24171efc379f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.344388 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rh4\" (UniqueName: \"kubernetes.io/projected/32057f35-2096-4a60-be4d-24171efc379f-kube-api-access-85rh4\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n8x75\" (UID: \"32057f35-2096-4a60-be4d-24171efc379f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.445309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rh4\" (UniqueName: \"kubernetes.io/projected/32057f35-2096-4a60-be4d-24171efc379f-kube-api-access-85rh4\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n8x75\" (UID: \"32057f35-2096-4a60-be4d-24171efc379f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.445367 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/32057f35-2096-4a60-be4d-24171efc379f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n8x75\" (UID: \"32057f35-2096-4a60-be4d-24171efc379f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.446595 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/32057f35-2096-4a60-be4d-24171efc379f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n8x75\" (UID: \"32057f35-2096-4a60-be4d-24171efc379f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.468079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rh4\" (UniqueName: \"kubernetes.io/projected/32057f35-2096-4a60-be4d-24171efc379f-kube-api-access-85rh4\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n8x75\" (UID: \"32057f35-2096-4a60-be4d-24171efc379f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:03 crc kubenswrapper[4809]: I1205 11:25:03.563201 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" Dec 05 11:25:04 crc kubenswrapper[4809]: I1205 11:25:04.022593 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75"] Dec 05 11:25:04 crc kubenswrapper[4809]: I1205 11:25:04.180533 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" event={"ID":"32057f35-2096-4a60-be4d-24171efc379f","Type":"ContainerStarted","Data":"df440bd69a8bb557510466152569a2ab7ca3c8aca5e9ac72ba1a322a3b87c3e3"} Dec 05 11:25:07 crc kubenswrapper[4809]: I1205 11:25:07.298254 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-pgs44" Dec 05 11:25:12 crc kubenswrapper[4809]: I1205 11:25:12.237823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" event={"ID":"32057f35-2096-4a60-be4d-24171efc379f","Type":"ContainerStarted","Data":"f2471d2dc17f87e4ea3d44538f866a8eb410f016bb847385f8f74f778e89dd1f"} Dec 05 11:25:12 crc kubenswrapper[4809]: I1205 11:25:12.263602 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n8x75" podStartSLOduration=1.352775208 podStartE2EDuration="9.263569894s" podCreationTimestamp="2025-12-05 11:25:03 +0000 UTC" firstStartedPulling="2025-12-05 11:25:04.029555893 +0000 UTC m=+999.420532451" lastFinishedPulling="2025-12-05 11:25:11.940350569 +0000 UTC m=+1007.331327137" observedRunningTime="2025-12-05 11:25:12.26226383 +0000 UTC m=+1007.653240428" watchObservedRunningTime="2025-12-05 11:25:12.263569894 +0000 UTC m=+1007.654546492" Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.046449 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.046541 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.046607 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.047490 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd6b669d62c0db11d4d46d14f9c5c9d47f99d27bb3324e58211f411bd95dcbe7"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.047601 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://bd6b669d62c0db11d4d46d14f9c5c9d47f99d27bb3324e58211f411bd95dcbe7" gracePeriod=600 Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.250925 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="bd6b669d62c0db11d4d46d14f9c5c9d47f99d27bb3324e58211f411bd95dcbe7" exitCode=0 Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.250953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"bd6b669d62c0db11d4d46d14f9c5c9d47f99d27bb3324e58211f411bd95dcbe7"} Dec 05 11:25:14 crc kubenswrapper[4809]: I1205 11:25:14.251677 4809 scope.go:117] "RemoveContainer" containerID="95559463623c60efa0b7c2c68dd34e50145b75a4e23fb8a25ef08e1782d3a372" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.179103 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-2jdvh"] Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.180103 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.182385 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.183107 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.184937 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-nrthw" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.205963 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-2jdvh"] Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.220502 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/10c849bc-151a-4e79-a288-ae9d082d6596-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-2jdvh\" (UID: \"10c849bc-151a-4e79-a288-ae9d082d6596\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.220707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmq9f\" (UniqueName: \"kubernetes.io/projected/10c849bc-151a-4e79-a288-ae9d082d6596-kube-api-access-lmq9f\") pod \"cert-manager-webhook-f4fb5df64-2jdvh\" (UID: \"10c849bc-151a-4e79-a288-ae9d082d6596\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.261897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"f84b335fda3d018d7628fbb24eef2ea4ed19d76e6c066be3637c91567a56e36e"} Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.322615 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmq9f\" (UniqueName: \"kubernetes.io/projected/10c849bc-151a-4e79-a288-ae9d082d6596-kube-api-access-lmq9f\") pod \"cert-manager-webhook-f4fb5df64-2jdvh\" (UID: \"10c849bc-151a-4e79-a288-ae9d082d6596\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.322787 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/10c849bc-151a-4e79-a288-ae9d082d6596-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-2jdvh\" (UID: \"10c849bc-151a-4e79-a288-ae9d082d6596\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.341052 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/10c849bc-151a-4e79-a288-ae9d082d6596-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-2jdvh\" (UID: \"10c849bc-151a-4e79-a288-ae9d082d6596\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.341662 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmq9f\" (UniqueName: \"kubernetes.io/projected/10c849bc-151a-4e79-a288-ae9d082d6596-kube-api-access-lmq9f\") pod \"cert-manager-webhook-f4fb5df64-2jdvh\" (UID: \"10c849bc-151a-4e79-a288-ae9d082d6596\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.496682 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:15 crc kubenswrapper[4809]: I1205 11:25:15.794336 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-2jdvh"] Dec 05 11:25:15 crc kubenswrapper[4809]: W1205 11:25:15.798767 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c849bc_151a_4e79_a288_ae9d082d6596.slice/crio-0608244289c1c40b90c0a5042dfc0ca391d320d320c491ba0c268bf187934c0d WatchSource:0}: Error finding container 0608244289c1c40b90c0a5042dfc0ca391d320d320c491ba0c268bf187934c0d: Status 404 returned error can't find the container with id 0608244289c1c40b90c0a5042dfc0ca391d320d320c491ba0c268bf187934c0d Dec 05 11:25:16 crc kubenswrapper[4809]: I1205 11:25:16.270675 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" event={"ID":"10c849bc-151a-4e79-a288-ae9d082d6596","Type":"ContainerStarted","Data":"0608244289c1c40b90c0a5042dfc0ca391d320d320c491ba0c268bf187934c0d"} Dec 05 11:25:18 crc kubenswrapper[4809]: I1205 11:25:18.884663 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt"] Dec 05 11:25:18 crc kubenswrapper[4809]: I1205 11:25:18.886068 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:18 crc kubenswrapper[4809]: I1205 11:25:18.895002 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-fz4tt" Dec 05 11:25:18 crc kubenswrapper[4809]: I1205 11:25:18.899176 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt"] Dec 05 11:25:18 crc kubenswrapper[4809]: I1205 11:25:18.984489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk6nb\" (UniqueName: \"kubernetes.io/projected/a46e32ba-4592-4ada-8289-d1329ae7e5df-kube-api-access-zk6nb\") pod \"cert-manager-cainjector-855d9ccff4-6n6jt\" (UID: \"a46e32ba-4592-4ada-8289-d1329ae7e5df\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:18 crc kubenswrapper[4809]: I1205 11:25:18.984544 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a46e32ba-4592-4ada-8289-d1329ae7e5df-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-6n6jt\" (UID: \"a46e32ba-4592-4ada-8289-d1329ae7e5df\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:19 crc kubenswrapper[4809]: I1205 11:25:19.085344 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk6nb\" (UniqueName: \"kubernetes.io/projected/a46e32ba-4592-4ada-8289-d1329ae7e5df-kube-api-access-zk6nb\") pod \"cert-manager-cainjector-855d9ccff4-6n6jt\" (UID: \"a46e32ba-4592-4ada-8289-d1329ae7e5df\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:19 crc kubenswrapper[4809]: I1205 11:25:19.085406 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a46e32ba-4592-4ada-8289-d1329ae7e5df-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-6n6jt\" (UID: \"a46e32ba-4592-4ada-8289-d1329ae7e5df\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:19 crc kubenswrapper[4809]: I1205 11:25:19.117539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk6nb\" (UniqueName: \"kubernetes.io/projected/a46e32ba-4592-4ada-8289-d1329ae7e5df-kube-api-access-zk6nb\") pod \"cert-manager-cainjector-855d9ccff4-6n6jt\" (UID: \"a46e32ba-4592-4ada-8289-d1329ae7e5df\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:19 crc kubenswrapper[4809]: I1205 11:25:19.120137 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a46e32ba-4592-4ada-8289-d1329ae7e5df-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-6n6jt\" (UID: \"a46e32ba-4592-4ada-8289-d1329ae7e5df\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:19 crc kubenswrapper[4809]: I1205 11:25:19.213730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" Dec 05 11:25:23 crc kubenswrapper[4809]: I1205 11:25:23.231739 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt"] Dec 05 11:25:23 crc kubenswrapper[4809]: I1205 11:25:23.317523 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" event={"ID":"10c849bc-151a-4e79-a288-ae9d082d6596","Type":"ContainerStarted","Data":"5722e9aa94552c034dc790dfe638c977285a0c98b51b86aca882977aafc29d03"} Dec 05 11:25:23 crc kubenswrapper[4809]: I1205 11:25:23.318132 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:23 crc kubenswrapper[4809]: I1205 11:25:23.320189 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" event={"ID":"a46e32ba-4592-4ada-8289-d1329ae7e5df","Type":"ContainerStarted","Data":"b806513c68209d6d36c21d56c1e415eaff86c7bafcb99301e982ee4e22f9e720"} Dec 05 11:25:23 crc kubenswrapper[4809]: I1205 11:25:23.335824 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" podStartSLOduration=1.202361813 podStartE2EDuration="8.335798115s" podCreationTimestamp="2025-12-05 11:25:15 +0000 UTC" firstStartedPulling="2025-12-05 11:25:15.800518241 +0000 UTC m=+1011.191494799" lastFinishedPulling="2025-12-05 11:25:22.933954553 +0000 UTC m=+1018.324931101" observedRunningTime="2025-12-05 11:25:23.330846903 +0000 UTC m=+1018.721823491" watchObservedRunningTime="2025-12-05 11:25:23.335798115 +0000 UTC m=+1018.726774703" Dec 05 11:25:24 crc kubenswrapper[4809]: I1205 11:25:24.328482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" event={"ID":"a46e32ba-4592-4ada-8289-d1329ae7e5df","Type":"ContainerStarted","Data":"234fe501e81b4c9766e91b7974a7048f0ac2dbf7f26c8ca0d591d3e6b0bb7243"} Dec 05 11:25:24 crc kubenswrapper[4809]: I1205 11:25:24.349983 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-6n6jt" podStartSLOduration=6.34996323 podStartE2EDuration="6.34996323s" podCreationTimestamp="2025-12-05 11:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:25:24.345666136 +0000 UTC m=+1019.736642704" watchObservedRunningTime="2025-12-05 11:25:24.34996323 +0000 UTC m=+1019.740939798" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.576264 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qcfpj"] Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.578136 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.585905 4809 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-966x9" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.599807 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vszq9\" (UniqueName: \"kubernetes.io/projected/32f2f5dc-9189-4cb7-9e85-e41acd2a9626-kube-api-access-vszq9\") pod \"cert-manager-86cb77c54b-qcfpj\" (UID: \"32f2f5dc-9189-4cb7-9e85-e41acd2a9626\") " pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.600060 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/32f2f5dc-9189-4cb7-9e85-e41acd2a9626-bound-sa-token\") pod \"cert-manager-86cb77c54b-qcfpj\" (UID: \"32f2f5dc-9189-4cb7-9e85-e41acd2a9626\") " pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.604555 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qcfpj"] Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.701679 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vszq9\" (UniqueName: \"kubernetes.io/projected/32f2f5dc-9189-4cb7-9e85-e41acd2a9626-kube-api-access-vszq9\") pod \"cert-manager-86cb77c54b-qcfpj\" (UID: \"32f2f5dc-9189-4cb7-9e85-e41acd2a9626\") " pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.702134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/32f2f5dc-9189-4cb7-9e85-e41acd2a9626-bound-sa-token\") pod \"cert-manager-86cb77c54b-qcfpj\" (UID: \"32f2f5dc-9189-4cb7-9e85-e41acd2a9626\") " pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.725888 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/32f2f5dc-9189-4cb7-9e85-e41acd2a9626-bound-sa-token\") pod \"cert-manager-86cb77c54b-qcfpj\" (UID: \"32f2f5dc-9189-4cb7-9e85-e41acd2a9626\") " pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.726363 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vszq9\" (UniqueName: \"kubernetes.io/projected/32f2f5dc-9189-4cb7-9e85-e41acd2a9626-kube-api-access-vszq9\") pod \"cert-manager-86cb77c54b-qcfpj\" (UID: \"32f2f5dc-9189-4cb7-9e85-e41acd2a9626\") " pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:26 crc kubenswrapper[4809]: I1205 11:25:26.909743 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-qcfpj" Dec 05 11:25:27 crc kubenswrapper[4809]: I1205 11:25:27.426732 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qcfpj"] Dec 05 11:25:27 crc kubenswrapper[4809]: W1205 11:25:27.429302 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32f2f5dc_9189_4cb7_9e85_e41acd2a9626.slice/crio-7185d226fc9e600127b4b0b3544ea4265fe7f79abee010b098b9f60478f8c35d WatchSource:0}: Error finding container 7185d226fc9e600127b4b0b3544ea4265fe7f79abee010b098b9f60478f8c35d: Status 404 returned error can't find the container with id 7185d226fc9e600127b4b0b3544ea4265fe7f79abee010b098b9f60478f8c35d Dec 05 11:25:28 crc kubenswrapper[4809]: I1205 11:25:28.408955 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-qcfpj" event={"ID":"32f2f5dc-9189-4cb7-9e85-e41acd2a9626","Type":"ContainerStarted","Data":"c36e3be21b6adc6ba6ee07312c05651a4b762a76ebe69e88a231d025d71e90a1"} Dec 05 11:25:28 crc kubenswrapper[4809]: I1205 11:25:28.409305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-qcfpj" event={"ID":"32f2f5dc-9189-4cb7-9e85-e41acd2a9626","Type":"ContainerStarted","Data":"7185d226fc9e600127b4b0b3544ea4265fe7f79abee010b098b9f60478f8c35d"} Dec 05 11:25:28 crc kubenswrapper[4809]: I1205 11:25:28.434749 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-qcfpj" podStartSLOduration=2.43472048 podStartE2EDuration="2.43472048s" podCreationTimestamp="2025-12-05 11:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:25:28.430056936 +0000 UTC m=+1023.821033494" watchObservedRunningTime="2025-12-05 11:25:28.43472048 +0000 UTC m=+1023.825697038" Dec 05 11:25:30 crc kubenswrapper[4809]: I1205 11:25:30.502856 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-2jdvh" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.657600 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r2dtx"] Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.659548 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r2dtx" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.662340 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.663158 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-m92bg" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.677559 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.678402 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r2dtx"] Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.725798 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvwc5\" (UniqueName: \"kubernetes.io/projected/5dc8d76f-822f-4b1e-9bb8-22aebba81584-kube-api-access-zvwc5\") pod \"openstack-operator-index-r2dtx\" (UID: \"5dc8d76f-822f-4b1e-9bb8-22aebba81584\") " pod="openstack-operators/openstack-operator-index-r2dtx" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.827751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvwc5\" (UniqueName: \"kubernetes.io/projected/5dc8d76f-822f-4b1e-9bb8-22aebba81584-kube-api-access-zvwc5\") pod \"openstack-operator-index-r2dtx\" (UID: \"5dc8d76f-822f-4b1e-9bb8-22aebba81584\") " pod="openstack-operators/openstack-operator-index-r2dtx" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.857727 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvwc5\" (UniqueName: \"kubernetes.io/projected/5dc8d76f-822f-4b1e-9bb8-22aebba81584-kube-api-access-zvwc5\") pod \"openstack-operator-index-r2dtx\" (UID: \"5dc8d76f-822f-4b1e-9bb8-22aebba81584\") " pod="openstack-operators/openstack-operator-index-r2dtx" Dec 05 11:25:33 crc kubenswrapper[4809]: I1205 11:25:33.982137 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r2dtx" Dec 05 11:25:34 crc kubenswrapper[4809]: I1205 11:25:34.200661 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r2dtx"] Dec 05 11:25:34 crc kubenswrapper[4809]: W1205 11:25:34.211566 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5dc8d76f_822f_4b1e_9bb8_22aebba81584.slice/crio-45d66768df5784defa9bead2625d68fc04452e65978d4aa46967ee052d579921 WatchSource:0}: Error finding container 45d66768df5784defa9bead2625d68fc04452e65978d4aa46967ee052d579921: Status 404 returned error can't find the container with id 45d66768df5784defa9bead2625d68fc04452e65978d4aa46967ee052d579921 Dec 05 11:25:34 crc kubenswrapper[4809]: I1205 11:25:34.451144 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r2dtx" event={"ID":"5dc8d76f-822f-4b1e-9bb8-22aebba81584","Type":"ContainerStarted","Data":"45d66768df5784defa9bead2625d68fc04452e65978d4aa46967ee052d579921"} Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.011497 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r2dtx"] Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.482842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r2dtx" event={"ID":"5dc8d76f-822f-4b1e-9bb8-22aebba81584","Type":"ContainerStarted","Data":"d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9"} Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.483006 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-r2dtx" podUID="5dc8d76f-822f-4b1e-9bb8-22aebba81584" containerName="registry-server" containerID="cri-o://d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9" gracePeriod=2 Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.501396 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r2dtx" podStartSLOduration=1.530763861 podStartE2EDuration="4.501372882s" podCreationTimestamp="2025-12-05 11:25:33 +0000 UTC" firstStartedPulling="2025-12-05 11:25:34.212727033 +0000 UTC m=+1029.603703591" lastFinishedPulling="2025-12-05 11:25:37.183336064 +0000 UTC m=+1032.574312612" observedRunningTime="2025-12-05 11:25:37.498522337 +0000 UTC m=+1032.889498915" watchObservedRunningTime="2025-12-05 11:25:37.501372882 +0000 UTC m=+1032.892349450" Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.637719 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-md2wn"] Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.639087 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.646824 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-md2wn"] Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.688249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rthq5\" (UniqueName: \"kubernetes.io/projected/d8735bdd-709f-4c81-91a3-3ebca48c0d30-kube-api-access-rthq5\") pod \"openstack-operator-index-md2wn\" (UID: \"d8735bdd-709f-4c81-91a3-3ebca48c0d30\") " pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.795263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rthq5\" (UniqueName: \"kubernetes.io/projected/d8735bdd-709f-4c81-91a3-3ebca48c0d30-kube-api-access-rthq5\") pod \"openstack-operator-index-md2wn\" (UID: \"d8735bdd-709f-4c81-91a3-3ebca48c0d30\") " pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.842367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rthq5\" (UniqueName: \"kubernetes.io/projected/d8735bdd-709f-4c81-91a3-3ebca48c0d30-kube-api-access-rthq5\") pod \"openstack-operator-index-md2wn\" (UID: \"d8735bdd-709f-4c81-91a3-3ebca48c0d30\") " pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.887559 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r2dtx" Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.896076 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvwc5\" (UniqueName: \"kubernetes.io/projected/5dc8d76f-822f-4b1e-9bb8-22aebba81584-kube-api-access-zvwc5\") pod \"5dc8d76f-822f-4b1e-9bb8-22aebba81584\" (UID: \"5dc8d76f-822f-4b1e-9bb8-22aebba81584\") " Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.899997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc8d76f-822f-4b1e-9bb8-22aebba81584-kube-api-access-zvwc5" (OuterVolumeSpecName: "kube-api-access-zvwc5") pod "5dc8d76f-822f-4b1e-9bb8-22aebba81584" (UID: "5dc8d76f-822f-4b1e-9bb8-22aebba81584"). InnerVolumeSpecName "kube-api-access-zvwc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:25:37 crc kubenswrapper[4809]: I1205 11:25:37.970507 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.009411 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvwc5\" (UniqueName: \"kubernetes.io/projected/5dc8d76f-822f-4b1e-9bb8-22aebba81584-kube-api-access-zvwc5\") on node \"crc\" DevicePath \"\"" Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.407402 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-md2wn"] Dec 05 11:25:38 crc kubenswrapper[4809]: W1205 11:25:38.418061 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8735bdd_709f_4c81_91a3_3ebca48c0d30.slice/crio-a1faa5179a72c0966ce00a23e85f28a309d4ee86cd02913e28e2390d3868bfb9 WatchSource:0}: Error finding container a1faa5179a72c0966ce00a23e85f28a309d4ee86cd02913e28e2390d3868bfb9: Status 404 returned error can't find the container with id a1faa5179a72c0966ce00a23e85f28a309d4ee86cd02913e28e2390d3868bfb9 Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.489913 4809 generic.go:334] "Generic (PLEG): container finished" podID="5dc8d76f-822f-4b1e-9bb8-22aebba81584" containerID="d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9" exitCode=0 Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.489956 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r2dtx" Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.489990 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r2dtx" event={"ID":"5dc8d76f-822f-4b1e-9bb8-22aebba81584","Type":"ContainerDied","Data":"d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9"} Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.490020 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r2dtx" event={"ID":"5dc8d76f-822f-4b1e-9bb8-22aebba81584","Type":"ContainerDied","Data":"45d66768df5784defa9bead2625d68fc04452e65978d4aa46967ee052d579921"} Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.490041 4809 scope.go:117] "RemoveContainer" containerID="d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9" Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.491618 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-md2wn" event={"ID":"d8735bdd-709f-4c81-91a3-3ebca48c0d30","Type":"ContainerStarted","Data":"a1faa5179a72c0966ce00a23e85f28a309d4ee86cd02913e28e2390d3868bfb9"} Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.524056 4809 scope.go:117] "RemoveContainer" containerID="d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9" Dec 05 11:25:38 crc kubenswrapper[4809]: E1205 11:25:38.524716 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9\": container with ID starting with d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9 not found: ID does not exist" containerID="d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9" Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.524771 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9"} err="failed to get container status \"d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9\": rpc error: code = NotFound desc = could not find container \"d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9\": container with ID starting with d32ca1c1d9a1f51ae84462b0d364742f18e4671d609af13468ca7df63e0a7fc9 not found: ID does not exist" Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.554224 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r2dtx"] Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.557494 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-r2dtx"] Dec 05 11:25:38 crc kubenswrapper[4809]: I1205 11:25:38.882901 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dc8d76f-822f-4b1e-9bb8-22aebba81584" path="/var/lib/kubelet/pods/5dc8d76f-822f-4b1e-9bb8-22aebba81584/volumes" Dec 05 11:25:39 crc kubenswrapper[4809]: I1205 11:25:39.504171 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-md2wn" event={"ID":"d8735bdd-709f-4c81-91a3-3ebca48c0d30","Type":"ContainerStarted","Data":"cdcd0aa0b235f066444a2b2fb2e4b411b566cd40bb70199894ee0e53b506e86d"} Dec 05 11:25:39 crc kubenswrapper[4809]: I1205 11:25:39.533273 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-md2wn" podStartSLOduration=2.490470084 podStartE2EDuration="2.533234718s" podCreationTimestamp="2025-12-05 11:25:37 +0000 UTC" firstStartedPulling="2025-12-05 11:25:38.422558812 +0000 UTC m=+1033.813535380" lastFinishedPulling="2025-12-05 11:25:38.465323456 +0000 UTC m=+1033.856300014" observedRunningTime="2025-12-05 11:25:39.526254682 +0000 UTC m=+1034.917231240" watchObservedRunningTime="2025-12-05 11:25:39.533234718 +0000 UTC m=+1034.924211316" Dec 05 11:25:47 crc kubenswrapper[4809]: I1205 11:25:47.970677 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:47 crc kubenswrapper[4809]: I1205 11:25:47.974004 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:48 crc kubenswrapper[4809]: I1205 11:25:48.015776 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:48 crc kubenswrapper[4809]: I1205 11:25:48.619884 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-md2wn" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.267491 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb"] Dec 05 11:25:49 crc kubenswrapper[4809]: E1205 11:25:49.267860 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc8d76f-822f-4b1e-9bb8-22aebba81584" containerName="registry-server" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.267876 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc8d76f-822f-4b1e-9bb8-22aebba81584" containerName="registry-server" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.268019 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc8d76f-822f-4b1e-9bb8-22aebba81584" containerName="registry-server" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.269300 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.271563 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2tqqr" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.293123 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb"] Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.470422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pb7g\" (UniqueName: \"kubernetes.io/projected/3a3db807-0a5b-4fbb-ad41-86857e3ec884-kube-api-access-9pb7g\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.470495 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-bundle\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.470533 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-util\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.572155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pb7g\" (UniqueName: \"kubernetes.io/projected/3a3db807-0a5b-4fbb-ad41-86857e3ec884-kube-api-access-9pb7g\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.572280 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-bundle\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.572349 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-util\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.572978 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-util\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.573139 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-bundle\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.604291 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pb7g\" (UniqueName: \"kubernetes.io/projected/3a3db807-0a5b-4fbb-ad41-86857e3ec884-kube-api-access-9pb7g\") pod \"e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:49 crc kubenswrapper[4809]: I1205 11:25:49.894409 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:50 crc kubenswrapper[4809]: I1205 11:25:50.320368 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb"] Dec 05 11:25:51 crc kubenswrapper[4809]: I1205 11:25:51.615374 4809 generic.go:334] "Generic (PLEG): container finished" podID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerID="240d6bc72e2e6232ac74d46807c4192fd21ada372a56b9dfb89fe5e03e44968f" exitCode=0 Dec 05 11:25:51 crc kubenswrapper[4809]: I1205 11:25:51.615471 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" event={"ID":"3a3db807-0a5b-4fbb-ad41-86857e3ec884","Type":"ContainerDied","Data":"240d6bc72e2e6232ac74d46807c4192fd21ada372a56b9dfb89fe5e03e44968f"} Dec 05 11:25:51 crc kubenswrapper[4809]: I1205 11:25:51.615936 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" event={"ID":"3a3db807-0a5b-4fbb-ad41-86857e3ec884","Type":"ContainerStarted","Data":"6ddac87c703a62c24ba7e4010ed567465ce12fc1110bc37dce399e6f37dadaf6"} Dec 05 11:25:52 crc kubenswrapper[4809]: I1205 11:25:52.628170 4809 generic.go:334] "Generic (PLEG): container finished" podID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerID="78dec2ff08c9f216b547f0985aad03b23657530cff43203a94f3ffd4fe0590f1" exitCode=0 Dec 05 11:25:52 crc kubenswrapper[4809]: I1205 11:25:52.628221 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" event={"ID":"3a3db807-0a5b-4fbb-ad41-86857e3ec884","Type":"ContainerDied","Data":"78dec2ff08c9f216b547f0985aad03b23657530cff43203a94f3ffd4fe0590f1"} Dec 05 11:25:53 crc kubenswrapper[4809]: I1205 11:25:53.635784 4809 generic.go:334] "Generic (PLEG): container finished" podID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerID="73a680825d276fedd50f0f4ea817f57e1bbaa22ab3b98ef84f8e030beb203343" exitCode=0 Dec 05 11:25:53 crc kubenswrapper[4809]: I1205 11:25:53.635847 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" event={"ID":"3a3db807-0a5b-4fbb-ad41-86857e3ec884","Type":"ContainerDied","Data":"73a680825d276fedd50f0f4ea817f57e1bbaa22ab3b98ef84f8e030beb203343"} Dec 05 11:25:54 crc kubenswrapper[4809]: I1205 11:25:54.894808 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.057446 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pb7g\" (UniqueName: \"kubernetes.io/projected/3a3db807-0a5b-4fbb-ad41-86857e3ec884-kube-api-access-9pb7g\") pod \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.057508 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-bundle\") pod \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.057557 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-util\") pod \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\" (UID: \"3a3db807-0a5b-4fbb-ad41-86857e3ec884\") " Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.058412 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-bundle" (OuterVolumeSpecName: "bundle") pod "3a3db807-0a5b-4fbb-ad41-86857e3ec884" (UID: "3a3db807-0a5b-4fbb-ad41-86857e3ec884"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.067840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3db807-0a5b-4fbb-ad41-86857e3ec884-kube-api-access-9pb7g" (OuterVolumeSpecName: "kube-api-access-9pb7g") pod "3a3db807-0a5b-4fbb-ad41-86857e3ec884" (UID: "3a3db807-0a5b-4fbb-ad41-86857e3ec884"). InnerVolumeSpecName "kube-api-access-9pb7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.086193 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-util" (OuterVolumeSpecName: "util") pod "3a3db807-0a5b-4fbb-ad41-86857e3ec884" (UID: "3a3db807-0a5b-4fbb-ad41-86857e3ec884"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.159579 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-util\") on node \"crc\" DevicePath \"\"" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.159617 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pb7g\" (UniqueName: \"kubernetes.io/projected/3a3db807-0a5b-4fbb-ad41-86857e3ec884-kube-api-access-9pb7g\") on node \"crc\" DevicePath \"\"" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.159647 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3a3db807-0a5b-4fbb-ad41-86857e3ec884-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.651534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" event={"ID":"3a3db807-0a5b-4fbb-ad41-86857e3ec884","Type":"ContainerDied","Data":"6ddac87c703a62c24ba7e4010ed567465ce12fc1110bc37dce399e6f37dadaf6"} Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.651584 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb" Dec 05 11:25:55 crc kubenswrapper[4809]: I1205 11:25:55.651604 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ddac87c703a62c24ba7e4010ed567465ce12fc1110bc37dce399e6f37dadaf6" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.154309 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4"] Dec 05 11:26:02 crc kubenswrapper[4809]: E1205 11:26:02.155440 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerName="pull" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.155456 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerName="pull" Dec 05 11:26:02 crc kubenswrapper[4809]: E1205 11:26:02.155480 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerName="util" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.155486 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerName="util" Dec 05 11:26:02 crc kubenswrapper[4809]: E1205 11:26:02.155504 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerName="extract" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.155511 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerName="extract" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.155659 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3db807-0a5b-4fbb-ad41-86857e3ec884" containerName="extract" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.156221 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.159326 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-lj2cs" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.186066 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4"] Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.251702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr62q\" (UniqueName: \"kubernetes.io/projected/9d9c431e-1776-4dbc-b5be-5218c3117128-kube-api-access-cr62q\") pod \"openstack-operator-controller-operator-6767b55986-hvln4\" (UID: \"9d9c431e-1776-4dbc-b5be-5218c3117128\") " pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.353441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr62q\" (UniqueName: \"kubernetes.io/projected/9d9c431e-1776-4dbc-b5be-5218c3117128-kube-api-access-cr62q\") pod \"openstack-operator-controller-operator-6767b55986-hvln4\" (UID: \"9d9c431e-1776-4dbc-b5be-5218c3117128\") " pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.387394 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr62q\" (UniqueName: \"kubernetes.io/projected/9d9c431e-1776-4dbc-b5be-5218c3117128-kube-api-access-cr62q\") pod \"openstack-operator-controller-operator-6767b55986-hvln4\" (UID: \"9d9c431e-1776-4dbc-b5be-5218c3117128\") " pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.475971 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" Dec 05 11:26:02 crc kubenswrapper[4809]: I1205 11:26:02.899624 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4"] Dec 05 11:26:03 crc kubenswrapper[4809]: I1205 11:26:03.700908 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" event={"ID":"9d9c431e-1776-4dbc-b5be-5218c3117128","Type":"ContainerStarted","Data":"d724bed76554b1dd74f31f4844004312c027e963291c2ed661f391dc42ae0649"} Dec 05 11:26:07 crc kubenswrapper[4809]: I1205 11:26:07.738460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" event={"ID":"9d9c431e-1776-4dbc-b5be-5218c3117128","Type":"ContainerStarted","Data":"8533ba9cbbb0866b4f5664f0a9a335b4ccf161c03623c66049cef83f87aa1a48"} Dec 05 11:26:07 crc kubenswrapper[4809]: I1205 11:26:07.739905 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" Dec 05 11:26:07 crc kubenswrapper[4809]: I1205 11:26:07.772128 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" podStartSLOduration=1.855526867 podStartE2EDuration="5.772102255s" podCreationTimestamp="2025-12-05 11:26:02 +0000 UTC" firstStartedPulling="2025-12-05 11:26:02.91815656 +0000 UTC m=+1058.309133118" lastFinishedPulling="2025-12-05 11:26:06.834731938 +0000 UTC m=+1062.225708506" observedRunningTime="2025-12-05 11:26:07.76896261 +0000 UTC m=+1063.159939178" watchObservedRunningTime="2025-12-05 11:26:07.772102255 +0000 UTC m=+1063.163078853" Dec 05 11:26:12 crc kubenswrapper[4809]: I1205 11:26:12.479689 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6767b55986-hvln4" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.768064 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.769317 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.771310 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-j8xjm" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.786504 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.787837 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.789953 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-8n7vc" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.793830 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.799286 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.800184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.801713 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-lw8lj" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.808681 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.816803 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.817584 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.817676 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.818596 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bn5s\" (UniqueName: \"kubernetes.io/projected/adae4108-5380-4358-b8bd-44ee6bb554af-kube-api-access-2bn5s\") pod \"designate-operator-controller-manager-78b4bc895b-cjt5m\" (UID: \"adae4108-5380-4358-b8bd-44ee6bb554af\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.818714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7868\" (UniqueName: \"kubernetes.io/projected/88757511-54a6-45b0-8bd8-90dddf1cae9f-kube-api-access-j7868\") pod \"glance-operator-controller-manager-77987cd8cd-plqqf\" (UID: \"88757511-54a6-45b0-8bd8-90dddf1cae9f\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.818759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqch8\" (UniqueName: \"kubernetes.io/projected/f14d52a2-20e5-429a-8285-82ff53430a63-kube-api-access-nqch8\") pod \"barbican-operator-controller-manager-7d9dfd778-cc22d\" (UID: \"f14d52a2-20e5-429a-8285-82ff53430a63\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.818790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4ccd\" (UniqueName: \"kubernetes.io/projected/ad1e2739-6610-4b87-aec7-89ca78575409-kube-api-access-p4ccd\") pod \"cinder-operator-controller-manager-859b6ccc6-nmg6x\" (UID: \"ad1e2739-6610-4b87-aec7-89ca78575409\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.819098 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gcldr" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.843176 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.877720 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.898980 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.901542 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-nzcq4" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.905201 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.916038 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.917137 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.919690 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqch8\" (UniqueName: \"kubernetes.io/projected/f14d52a2-20e5-429a-8285-82ff53430a63-kube-api-access-nqch8\") pod \"barbican-operator-controller-manager-7d9dfd778-cc22d\" (UID: \"f14d52a2-20e5-429a-8285-82ff53430a63\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.919737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rc9m\" (UniqueName: \"kubernetes.io/projected/84c8e973-2683-4917-bea9-a7f44eadf397-kube-api-access-9rc9m\") pod \"horizon-operator-controller-manager-68c6d99b8f-57hcg\" (UID: \"84c8e973-2683-4917-bea9-a7f44eadf397\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.919760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4ccd\" (UniqueName: \"kubernetes.io/projected/ad1e2739-6610-4b87-aec7-89ca78575409-kube-api-access-p4ccd\") pod \"cinder-operator-controller-manager-859b6ccc6-nmg6x\" (UID: \"ad1e2739-6610-4b87-aec7-89ca78575409\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.919817 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bn5s\" (UniqueName: \"kubernetes.io/projected/adae4108-5380-4358-b8bd-44ee6bb554af-kube-api-access-2bn5s\") pod \"designate-operator-controller-manager-78b4bc895b-cjt5m\" (UID: \"adae4108-5380-4358-b8bd-44ee6bb554af\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.919847 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7868\" (UniqueName: \"kubernetes.io/projected/88757511-54a6-45b0-8bd8-90dddf1cae9f-kube-api-access-j7868\") pod \"glance-operator-controller-manager-77987cd8cd-plqqf\" (UID: \"88757511-54a6-45b0-8bd8-90dddf1cae9f\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.924844 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-f8cwx" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.927098 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.942645 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bn5s\" (UniqueName: \"kubernetes.io/projected/adae4108-5380-4358-b8bd-44ee6bb554af-kube-api-access-2bn5s\") pod \"designate-operator-controller-manager-78b4bc895b-cjt5m\" (UID: \"adae4108-5380-4358-b8bd-44ee6bb554af\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.943056 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqch8\" (UniqueName: \"kubernetes.io/projected/f14d52a2-20e5-429a-8285-82ff53430a63-kube-api-access-nqch8\") pod \"barbican-operator-controller-manager-7d9dfd778-cc22d\" (UID: \"f14d52a2-20e5-429a-8285-82ff53430a63\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.947191 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.948143 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.950057 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-cjbm8" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.951715 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.952830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.957015 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-q7mx7" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.957237 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.960157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4ccd\" (UniqueName: \"kubernetes.io/projected/ad1e2739-6610-4b87-aec7-89ca78575409-kube-api-access-p4ccd\") pod \"cinder-operator-controller-manager-859b6ccc6-nmg6x\" (UID: \"ad1e2739-6610-4b87-aec7-89ca78575409\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.962103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7868\" (UniqueName: \"kubernetes.io/projected/88757511-54a6-45b0-8bd8-90dddf1cae9f-kube-api-access-j7868\") pod \"glance-operator-controller-manager-77987cd8cd-plqqf\" (UID: \"88757511-54a6-45b0-8bd8-90dddf1cae9f\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.963675 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.974448 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.984964 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf"] Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.990465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.995130 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-54nfs" Dec 05 11:26:31 crc kubenswrapper[4809]: I1205 11:26:31.995321 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.020701 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.021704 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.025427 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-g52qk" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.027174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slp92\" (UniqueName: \"kubernetes.io/projected/9eec44c6-6e4d-4945-afef-ed95846e7b50-kube-api-access-slp92\") pod \"heat-operator-controller-manager-5f64f6f8bb-9q4zp\" (UID: \"9eec44c6-6e4d-4945-afef-ed95846e7b50\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.027218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqdjv\" (UniqueName: \"kubernetes.io/projected/ab4b415e-554a-4e1b-a805-d628772bc6bc-kube-api-access-mqdjv\") pod \"manila-operator-controller-manager-7c79b5df47-57zmz\" (UID: \"ab4b415e-554a-4e1b-a805-d628772bc6bc\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.027281 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rc9m\" (UniqueName: \"kubernetes.io/projected/84c8e973-2683-4917-bea9-a7f44eadf397-kube-api-access-9rc9m\") pod \"horizon-operator-controller-manager-68c6d99b8f-57hcg\" (UID: \"84c8e973-2683-4917-bea9-a7f44eadf397\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.036911 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.040955 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.041854 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.044756 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-cf2pz" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.062291 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.063614 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.074542 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rc9m\" (UniqueName: \"kubernetes.io/projected/84c8e973-2683-4917-bea9-a7f44eadf397-kube-api-access-9rc9m\") pod \"horizon-operator-controller-manager-68c6d99b8f-57hcg\" (UID: \"84c8e973-2683-4917-bea9-a7f44eadf397\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.079587 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-cw7x4" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.109732 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.123114 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.133964 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slp92\" (UniqueName: \"kubernetes.io/projected/9eec44c6-6e4d-4945-afef-ed95846e7b50-kube-api-access-slp92\") pod \"heat-operator-controller-manager-5f64f6f8bb-9q4zp\" (UID: \"9eec44c6-6e4d-4945-afef-ed95846e7b50\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.134016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqdjv\" (UniqueName: \"kubernetes.io/projected/ab4b415e-554a-4e1b-a805-d628772bc6bc-kube-api-access-mqdjv\") pod \"manila-operator-controller-manager-7c79b5df47-57zmz\" (UID: \"ab4b415e-554a-4e1b-a805-d628772bc6bc\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.134086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.134112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cwbf\" (UniqueName: \"kubernetes.io/projected/06bcaa51-2671-458b-9cd0-a4c73455b02a-kube-api-access-2cwbf\") pod \"ironic-operator-controller-manager-6c548fd776-jgxpj\" (UID: \"06bcaa51-2671-458b-9cd0-a4c73455b02a\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.134173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-926sq\" (UniqueName: \"kubernetes.io/projected/e664f482-5587-4ad2-a35e-0fe439d6de39-kube-api-access-926sq\") pod \"keystone-operator-controller-manager-7765d96ddf-jdbkf\" (UID: \"e664f482-5587-4ad2-a35e-0fe439d6de39\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.134205 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crs6h\" (UniqueName: \"kubernetes.io/projected/6167f262-8efb-43c5-b065-4a2de04a183f-kube-api-access-crs6h\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.135272 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.140546 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.142597 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.146120 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-snvwk" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.146397 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.148014 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.153153 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-pjsrc" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.153328 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.160079 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.167267 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.183508 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slp92\" (UniqueName: \"kubernetes.io/projected/9eec44c6-6e4d-4945-afef-ed95846e7b50-kube-api-access-slp92\") pod \"heat-operator-controller-manager-5f64f6f8bb-9q4zp\" (UID: \"9eec44c6-6e4d-4945-afef-ed95846e7b50\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.206031 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.221415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqdjv\" (UniqueName: \"kubernetes.io/projected/ab4b415e-554a-4e1b-a805-d628772bc6bc-kube-api-access-mqdjv\") pod \"manila-operator-controller-manager-7c79b5df47-57zmz\" (UID: \"ab4b415e-554a-4e1b-a805-d628772bc6bc\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.221809 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.227304 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.235320 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4j7p\" (UniqueName: \"kubernetes.io/projected/990aee67-4ef0-4a34-b228-d511085caf18-kube-api-access-s4j7p\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-dmgm2\" (UID: \"990aee67-4ef0-4a34-b228-d511085caf18\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.235401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.235440 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cwbf\" (UniqueName: \"kubernetes.io/projected/06bcaa51-2671-458b-9cd0-a4c73455b02a-kube-api-access-2cwbf\") pod \"ironic-operator-controller-manager-6c548fd776-jgxpj\" (UID: \"06bcaa51-2671-458b-9cd0-a4c73455b02a\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.235506 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9vld\" (UniqueName: \"kubernetes.io/projected/2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15-kube-api-access-h9vld\") pod \"mariadb-operator-controller-manager-56bbcc9d85-7z7pv\" (UID: \"2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.235532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-926sq\" (UniqueName: \"kubernetes.io/projected/e664f482-5587-4ad2-a35e-0fe439d6de39-kube-api-access-926sq\") pod \"keystone-operator-controller-manager-7765d96ddf-jdbkf\" (UID: \"e664f482-5587-4ad2-a35e-0fe439d6de39\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.235565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crs6h\" (UniqueName: \"kubernetes.io/projected/6167f262-8efb-43c5-b065-4a2de04a183f-kube-api-access-crs6h\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:32 crc kubenswrapper[4809]: E1205 11:26:32.236117 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:32 crc kubenswrapper[4809]: E1205 11:26:32.236334 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert podName:6167f262-8efb-43c5-b065-4a2de04a183f nodeName:}" failed. No retries permitted until 2025-12-05 11:26:32.736306264 +0000 UTC m=+1088.127282812 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert") pod "infra-operator-controller-manager-57548d458d-k8cxw" (UID: "6167f262-8efb-43c5-b065-4a2de04a183f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.239765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.245704 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.247073 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.257649 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.258154 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-d7hck" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.291356 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-926sq\" (UniqueName: \"kubernetes.io/projected/e664f482-5587-4ad2-a35e-0fe439d6de39-kube-api-access-926sq\") pod \"keystone-operator-controller-manager-7765d96ddf-jdbkf\" (UID: \"e664f482-5587-4ad2-a35e-0fe439d6de39\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.302138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crs6h\" (UniqueName: \"kubernetes.io/projected/6167f262-8efb-43c5-b065-4a2de04a183f-kube-api-access-crs6h\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.303700 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cwbf\" (UniqueName: \"kubernetes.io/projected/06bcaa51-2671-458b-9cd0-a4c73455b02a-kube-api-access-2cwbf\") pod \"ironic-operator-controller-manager-6c548fd776-jgxpj\" (UID: \"06bcaa51-2671-458b-9cd0-a4c73455b02a\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.309049 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.335492 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.336979 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4lww\" (UniqueName: \"kubernetes.io/projected/094859bf-906c-4efa-b894-29db3182659b-kube-api-access-x4lww\") pod \"nova-operator-controller-manager-697bc559fc-c27fb\" (UID: \"094859bf-906c-4efa-b894-29db3182659b\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.337052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rpbq\" (UniqueName: \"kubernetes.io/projected/d47fc62a-29c5-4ebc-a3b4-45d6ce997907-kube-api-access-7rpbq\") pod \"octavia-operator-controller-manager-998648c74-hx8w2\" (UID: \"d47fc62a-29c5-4ebc-a3b4-45d6ce997907\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.337100 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4j7p\" (UniqueName: \"kubernetes.io/projected/990aee67-4ef0-4a34-b228-d511085caf18-kube-api-access-s4j7p\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-dmgm2\" (UID: \"990aee67-4ef0-4a34-b228-d511085caf18\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.337200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9vld\" (UniqueName: \"kubernetes.io/projected/2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15-kube-api-access-h9vld\") pod \"mariadb-operator-controller-manager-56bbcc9d85-7z7pv\" (UID: \"2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.337944 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.342205 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-49t9f" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.348732 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.372051 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.401921 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4j7p\" (UniqueName: \"kubernetes.io/projected/990aee67-4ef0-4a34-b228-d511085caf18-kube-api-access-s4j7p\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-dmgm2\" (UID: \"990aee67-4ef0-4a34-b228-d511085caf18\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.402255 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.403779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9vld\" (UniqueName: \"kubernetes.io/projected/2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15-kube-api-access-h9vld\") pod \"mariadb-operator-controller-manager-56bbcc9d85-7z7pv\" (UID: \"2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.413739 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.417229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.433659 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.442789 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8bqt\" (UniqueName: \"kubernetes.io/projected/53e22095-69df-44e2-aaea-7a6fbbbfce6c-kube-api-access-t8bqt\") pod \"ovn-operator-controller-manager-b6456fdb6-mh2r6\" (UID: \"53e22095-69df-44e2-aaea-7a6fbbbfce6c\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.442893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nv86\" (UniqueName: \"kubernetes.io/projected/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-kube-api-access-9nv86\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.442984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4lww\" (UniqueName: \"kubernetes.io/projected/094859bf-906c-4efa-b894-29db3182659b-kube-api-access-x4lww\") pod \"nova-operator-controller-manager-697bc559fc-c27fb\" (UID: \"094859bf-906c-4efa-b894-29db3182659b\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.443047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rpbq\" (UniqueName: \"kubernetes.io/projected/d47fc62a-29c5-4ebc-a3b4-45d6ce997907-kube-api-access-7rpbq\") pod \"octavia-operator-controller-manager-998648c74-hx8w2\" (UID: \"d47fc62a-29c5-4ebc-a3b4-45d6ce997907\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.443088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.486033 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4lww\" (UniqueName: \"kubernetes.io/projected/094859bf-906c-4efa-b894-29db3182659b-kube-api-access-x4lww\") pod \"nova-operator-controller-manager-697bc559fc-c27fb\" (UID: \"094859bf-906c-4efa-b894-29db3182659b\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.489271 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.491131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.494771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rpbq\" (UniqueName: \"kubernetes.io/projected/d47fc62a-29c5-4ebc-a3b4-45d6ce997907-kube-api-access-7rpbq\") pod \"octavia-operator-controller-manager-998648c74-hx8w2\" (UID: \"d47fc62a-29c5-4ebc-a3b4-45d6ce997907\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.495862 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-tdrw5" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.506288 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.519836 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.523826 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-jtvtc" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.546197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.546306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8bqt\" (UniqueName: \"kubernetes.io/projected/53e22095-69df-44e2-aaea-7a6fbbbfce6c-kube-api-access-t8bqt\") pod \"ovn-operator-controller-manager-b6456fdb6-mh2r6\" (UID: \"53e22095-69df-44e2-aaea-7a6fbbbfce6c\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.546349 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nv86\" (UniqueName: \"kubernetes.io/projected/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-kube-api-access-9nv86\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:32 crc kubenswrapper[4809]: E1205 11:26:32.547058 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:32 crc kubenswrapper[4809]: E1205 11:26:32.547160 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert podName:7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:33.047133449 +0000 UTC m=+1088.438110187 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" (UID: "7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.551354 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.557173 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.577423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nv86\" (UniqueName: \"kubernetes.io/projected/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-kube-api-access-9nv86\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.588175 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8bqt\" (UniqueName: \"kubernetes.io/projected/53e22095-69df-44e2-aaea-7a6fbbbfce6c-kube-api-access-t8bqt\") pod \"ovn-operator-controller-manager-b6456fdb6-mh2r6\" (UID: \"53e22095-69df-44e2-aaea-7a6fbbbfce6c\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.603735 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.621248 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.644362 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.646013 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.649432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw2nc\" (UniqueName: \"kubernetes.io/projected/8728b108-ea52-4024-bbba-d1985b5d9795-kube-api-access-qw2nc\") pod \"placement-operator-controller-manager-78f8948974-5sz6t\" (UID: \"8728b108-ea52-4024-bbba-d1985b5d9795\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.649482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjz2d\" (UniqueName: \"kubernetes.io/projected/ac756b27-3bb1-4b8d-925c-15b41db189c8-kube-api-access-fjz2d\") pod \"swift-operator-controller-manager-5f8c65bbfc-9m782\" (UID: \"ac756b27-3bb1-4b8d-925c-15b41db189c8\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.652002 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-6m7ms" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.663397 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.680948 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-khmw4"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.682298 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.686256 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-qd8rq" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.688619 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-khmw4"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.693723 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.704599 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.706243 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.710329 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.711506 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-c6nbp" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.752316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hslbn\" (UniqueName: \"kubernetes.io/projected/f586c1d3-4a51-48ff-be7f-d229d2770c10-kube-api-access-hslbn\") pod \"telemetry-operator-controller-manager-76cc84c6bb-ctckv\" (UID: \"f586c1d3-4a51-48ff-be7f-d229d2770c10\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.752439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw2nc\" (UniqueName: \"kubernetes.io/projected/8728b108-ea52-4024-bbba-d1985b5d9795-kube-api-access-qw2nc\") pod \"placement-operator-controller-manager-78f8948974-5sz6t\" (UID: \"8728b108-ea52-4024-bbba-d1985b5d9795\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.752488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjz2d\" (UniqueName: \"kubernetes.io/projected/ac756b27-3bb1-4b8d-925c-15b41db189c8-kube-api-access-fjz2d\") pod \"swift-operator-controller-manager-5f8c65bbfc-9m782\" (UID: \"ac756b27-3bb1-4b8d-925c-15b41db189c8\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.752535 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:32 crc kubenswrapper[4809]: E1205 11:26:32.752682 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:32 crc kubenswrapper[4809]: E1205 11:26:32.752745 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert podName:6167f262-8efb-43c5-b065-4a2de04a183f nodeName:}" failed. No retries permitted until 2025-12-05 11:26:33.752728391 +0000 UTC m=+1089.143705019 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert") pod "infra-operator-controller-manager-57548d458d-k8cxw" (UID: "6167f262-8efb-43c5-b065-4a2de04a183f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.768247 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.769255 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.771168 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.771573 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.774073 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-58zn6" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.777059 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw2nc\" (UniqueName: \"kubernetes.io/projected/8728b108-ea52-4024-bbba-d1985b5d9795-kube-api-access-qw2nc\") pod \"placement-operator-controller-manager-78f8948974-5sz6t\" (UID: \"8728b108-ea52-4024-bbba-d1985b5d9795\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.782291 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.790425 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.791843 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.794078 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-xzhv5" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.797038 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.798798 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjz2d\" (UniqueName: \"kubernetes.io/projected/ac756b27-3bb1-4b8d-925c-15b41db189c8-kube-api-access-fjz2d\") pod \"swift-operator-controller-manager-5f8c65bbfc-9m782\" (UID: \"ac756b27-3bb1-4b8d-925c-15b41db189c8\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.838977 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.854958 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nl4j\" (UniqueName: \"kubernetes.io/projected/adcbada3-1e83-4051-a036-f65cce0eb095-kube-api-access-8nl4j\") pod \"test-operator-controller-manager-5854674fcc-khmw4\" (UID: \"adcbada3-1e83-4051-a036-f65cce0eb095\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.855070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hslbn\" (UniqueName: \"kubernetes.io/projected/f586c1d3-4a51-48ff-be7f-d229d2770c10-kube-api-access-hslbn\") pod \"telemetry-operator-controller-manager-76cc84c6bb-ctckv\" (UID: \"f586c1d3-4a51-48ff-be7f-d229d2770c10\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.855123 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7c4g\" (UniqueName: \"kubernetes.io/projected/a738af9f-fa2b-44ea-816a-135b351ef9d2-kube-api-access-w7c4g\") pod \"watcher-operator-controller-manager-769dc69bc-vf5hg\" (UID: \"a738af9f-fa2b-44ea-816a-135b351ef9d2\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.860131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.920558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hslbn\" (UniqueName: \"kubernetes.io/projected/f586c1d3-4a51-48ff-be7f-d229d2770c10-kube-api-access-hslbn\") pod \"telemetry-operator-controller-manager-76cc84c6bb-ctckv\" (UID: \"f586c1d3-4a51-48ff-be7f-d229d2770c10\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.956514 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nl4j\" (UniqueName: \"kubernetes.io/projected/adcbada3-1e83-4051-a036-f65cce0eb095-kube-api-access-8nl4j\") pod \"test-operator-controller-manager-5854674fcc-khmw4\" (UID: \"adcbada3-1e83-4051-a036-f65cce0eb095\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.956572 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7c4g\" (UniqueName: \"kubernetes.io/projected/a738af9f-fa2b-44ea-816a-135b351ef9d2-kube-api-access-w7c4g\") pod \"watcher-operator-controller-manager-769dc69bc-vf5hg\" (UID: \"a738af9f-fa2b-44ea-816a-135b351ef9d2\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.956655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.956699 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlf95\" (UniqueName: \"kubernetes.io/projected/69cacfb4-8a61-4bd4-849f-94fc24d85453-kube-api-access-mlf95\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.956738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.956814 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf62k\" (UniqueName: \"kubernetes.io/projected/8bc5d720-4c3a-487b-a58d-4dc4c5760d6e-kube-api-access-rf62k\") pod \"rabbitmq-cluster-operator-manager-668c99d594-w8rhm\" (UID: \"8bc5d720-4c3a-487b-a58d-4dc4c5760d6e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.969232 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d"] Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.984466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7c4g\" (UniqueName: \"kubernetes.io/projected/a738af9f-fa2b-44ea-816a-135b351ef9d2-kube-api-access-w7c4g\") pod \"watcher-operator-controller-manager-769dc69bc-vf5hg\" (UID: \"a738af9f-fa2b-44ea-816a-135b351ef9d2\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" Dec 05 11:26:32 crc kubenswrapper[4809]: I1205 11:26:32.985110 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nl4j\" (UniqueName: \"kubernetes.io/projected/adcbada3-1e83-4051-a036-f65cce0eb095-kube-api-access-8nl4j\") pod \"test-operator-controller-manager-5854674fcc-khmw4\" (UID: \"adcbada3-1e83-4051-a036-f65cce0eb095\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.047776 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.058210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf62k\" (UniqueName: \"kubernetes.io/projected/8bc5d720-4c3a-487b-a58d-4dc4c5760d6e-kube-api-access-rf62k\") pod \"rabbitmq-cluster-operator-manager-668c99d594-w8rhm\" (UID: \"8bc5d720-4c3a-487b-a58d-4dc4c5760d6e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.058416 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.058459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlf95\" (UniqueName: \"kubernetes.io/projected/69cacfb4-8a61-4bd4-849f-94fc24d85453-kube-api-access-mlf95\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.058487 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.058513 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.058676 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.058737 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:33.558717945 +0000 UTC m=+1088.949694503 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "metrics-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.059412 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.059464 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:33.559447365 +0000 UTC m=+1088.950423913 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.059969 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.060098 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert podName:7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:34.060071012 +0000 UTC m=+1089.451047570 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" (UID: "7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.091262 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlf95\" (UniqueName: \"kubernetes.io/projected/69cacfb4-8a61-4bd4-849f-94fc24d85453-kube-api-access-mlf95\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.093764 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf62k\" (UniqueName: \"kubernetes.io/projected/8bc5d720-4c3a-487b-a58d-4dc4c5760d6e-kube-api-access-rf62k\") pod \"rabbitmq-cluster-operator-manager-668c99d594-w8rhm\" (UID: \"8bc5d720-4c3a-487b-a58d-4dc4c5760d6e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.119953 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.140408 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.234005 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.261260 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.265987 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.277710 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m"] Dec 05 11:26:33 crc kubenswrapper[4809]: W1205 11:26:33.324210 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84c8e973_2683_4917_bea9_a7f44eadf397.slice/crio-c2c7cd3a498db2e673b4ccfd108e6ca47712631fbbb27db03ac13b4d80baae52 WatchSource:0}: Error finding container c2c7cd3a498db2e673b4ccfd108e6ca47712631fbbb27db03ac13b4d80baae52: Status 404 returned error can't find the container with id c2c7cd3a498db2e673b4ccfd108e6ca47712631fbbb27db03ac13b4d80baae52 Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.567850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.567908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.568023 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.568066 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:34.56805347 +0000 UTC m=+1089.959030028 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "metrics-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.568645 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.568674 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:34.568665876 +0000 UTC m=+1089.959642434 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.729027 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.743846 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2"] Dec 05 11:26:33 crc kubenswrapper[4809]: W1205 11:26:33.744222 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad1e2739_6610_4b87_aec7_89ca78575409.slice/crio-a47868f2a13d0398af30e4bbd1d9291b7bb581dc73ae092cf6c01a652a4a84e8 WatchSource:0}: Error finding container a47868f2a13d0398af30e4bbd1d9291b7bb581dc73ae092cf6c01a652a4a84e8: Status 404 returned error can't find the container with id a47868f2a13d0398af30e4bbd1d9291b7bb581dc73ae092cf6c01a652a4a84e8 Dec 05 11:26:33 crc kubenswrapper[4809]: W1205 11:26:33.745782 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eec44c6_6e4d_4945_afef_ed95846e7b50.slice/crio-a2043c289249bbee70c5b8f29a44cd5174d3412b0ff458afa47e0fed0209f761 WatchSource:0}: Error finding container a2043c289249bbee70c5b8f29a44cd5174d3412b0ff458afa47e0fed0209f761: Status 404 returned error can't find the container with id a2043c289249bbee70c5b8f29a44cd5174d3412b0ff458afa47e0fed0209f761 Dec 05 11:26:33 crc kubenswrapper[4809]: W1205 11:26:33.746170 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd47fc62a_29c5_4ebc_a3b4_45d6ce997907.slice/crio-98a22372bcf6e5c5e558da9ac43d00278c4d0a6fc320a5d2b0e251b5795aa2f2 WatchSource:0}: Error finding container 98a22372bcf6e5c5e558da9ac43d00278c4d0a6fc320a5d2b0e251b5795aa2f2: Status 404 returned error can't find the container with id 98a22372bcf6e5c5e558da9ac43d00278c4d0a6fc320a5d2b0e251b5795aa2f2 Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.759601 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.767997 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.771324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.771522 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: E1205 11:26:33.771665 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert podName:6167f262-8efb-43c5-b065-4a2de04a183f nodeName:}" failed. No retries permitted until 2025-12-05 11:26:35.771622857 +0000 UTC m=+1091.162599415 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert") pod "infra-operator-controller-manager-57548d458d-k8cxw" (UID: "6167f262-8efb-43c5-b065-4a2de04a183f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.774389 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.946056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" event={"ID":"84c8e973-2683-4917-bea9-a7f44eadf397","Type":"ContainerStarted","Data":"c2c7cd3a498db2e673b4ccfd108e6ca47712631fbbb27db03ac13b4d80baae52"} Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.950927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" event={"ID":"adae4108-5380-4358-b8bd-44ee6bb554af","Type":"ContainerStarted","Data":"859226f280ff19ad688874c67cb843c3dc669db11de73370bec249727937992c"} Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.952429 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" event={"ID":"06bcaa51-2671-458b-9cd0-a4c73455b02a","Type":"ContainerStarted","Data":"d5ac7c9499fb3cc5440cd3e1a5fdfbc75fa1805360e4087827686e7675f4536a"} Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.953315 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" event={"ID":"88757511-54a6-45b0-8bd8-90dddf1cae9f","Type":"ContainerStarted","Data":"a51aeb5b199f9803aa99e89521ac020302f6bfef2f4c99c0f7381bd56f3b4178"} Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.963090 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.963207 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" event={"ID":"f14d52a2-20e5-429a-8285-82ff53430a63","Type":"ContainerStarted","Data":"4d99e59de708bd71f126030dc02b51de93b3878dfb3df6714932683fc7cd129a"} Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.971872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" event={"ID":"9eec44c6-6e4d-4945-afef-ed95846e7b50","Type":"ContainerStarted","Data":"a2043c289249bbee70c5b8f29a44cd5174d3412b0ff458afa47e0fed0209f761"} Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.978721 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.985033 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-khmw4"] Dec 05 11:26:33 crc kubenswrapper[4809]: W1205 11:26:33.990067 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bc5d720_4c3a_487b_a58d_4dc4c5760d6e.slice/crio-313269c0552077be4bc225ddcc08eebd7e2d0f7affc79def807bffd716e2c617 WatchSource:0}: Error finding container 313269c0552077be4bc225ddcc08eebd7e2d0f7affc79def807bffd716e2c617: Status 404 returned error can't find the container with id 313269c0552077be4bc225ddcc08eebd7e2d0f7affc79def807bffd716e2c617 Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.990820 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg"] Dec 05 11:26:33 crc kubenswrapper[4809]: I1205 11:26:33.997581 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" event={"ID":"d47fc62a-29c5-4ebc-a3b4-45d6ce997907","Type":"ContainerStarted","Data":"98a22372bcf6e5c5e558da9ac43d00278c4d0a6fc320a5d2b0e251b5795aa2f2"} Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.001275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" event={"ID":"ad1e2739-6610-4b87-aec7-89ca78575409","Type":"ContainerStarted","Data":"a47868f2a13d0398af30e4bbd1d9291b7bb581dc73ae092cf6c01a652a4a84e8"} Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.002445 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb"] Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.007222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" event={"ID":"53e22095-69df-44e2-aaea-7a6fbbbfce6c","Type":"ContainerStarted","Data":"0f711ce3577760ca3d01e37f880079ed78ca0d53469b2a754672c7df38846d97"} Dec 05 11:26:34 crc kubenswrapper[4809]: W1205 11:26:34.007782 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab4b415e_554a_4e1b_a805_d628772bc6bc.slice/crio-9f5262c545b79adf6a9bf904dc183f1168566b28830ead018756575aca7902e5 WatchSource:0}: Error finding container 9f5262c545b79adf6a9bf904dc183f1168566b28830ead018756575aca7902e5: Status 404 returned error can't find the container with id 9f5262c545b79adf6a9bf904dc183f1168566b28830ead018756575aca7902e5 Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.008262 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz"] Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.017758 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2"] Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.024481 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv"] Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.039833 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t"] Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.044300 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf"] Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.049799 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x4lww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-c27fb_openstack-operators(094859bf-906c-4efa-b894-29db3182659b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.052021 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8nl4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-khmw4_openstack-operators(adcbada3-1e83-4051-a036-f65cce0eb095): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.052927 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x4lww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-c27fb_openstack-operators(094859bf-906c-4efa-b894-29db3182659b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.053278 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qw2nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-5sz6t_openstack-operators(8728b108-ea52-4024-bbba-d1985b5d9795): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.054438 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" podUID="094859bf-906c-4efa-b894-29db3182659b" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.055920 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8nl4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-khmw4_openstack-operators(adcbada3-1e83-4051-a036-f65cce0eb095): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.058755 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" podUID="adcbada3-1e83-4051-a036-f65cce0eb095" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.058937 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qw2nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-5sz6t_openstack-operators(8728b108-ea52-4024-bbba-d1985b5d9795): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.060308 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" podUID="8728b108-ea52-4024-bbba-d1985b5d9795" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.064344 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-926sq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-jdbkf_openstack-operators(e664f482-5587-4ad2-a35e-0fe439d6de39): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.069130 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-926sq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-jdbkf_openstack-operators(e664f482-5587-4ad2-a35e-0fe439d6de39): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.070371 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" podUID="e664f482-5587-4ad2-a35e-0fe439d6de39" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.071345 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h9vld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-7z7pv_openstack-operators(2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.074946 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h9vld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-7z7pv_openstack-operators(2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.075567 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.075819 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.075869 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert podName:7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:36.075855223 +0000 UTC m=+1091.466831781 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" (UID: "7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.076220 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" podUID="2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15" Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.171443 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782"] Dec 05 11:26:34 crc kubenswrapper[4809]: W1205 11:26:34.182512 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac756b27_3bb1_4b8d_925c_15b41db189c8.slice/crio-0cb5714a39aa180081b2adc05e01ea08eda047116fd9308684c584768e04c9d4 WatchSource:0}: Error finding container 0cb5714a39aa180081b2adc05e01ea08eda047116fd9308684c584768e04c9d4: Status 404 returned error can't find the container with id 0cb5714a39aa180081b2adc05e01ea08eda047116fd9308684c584768e04c9d4 Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.589210 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:34 crc kubenswrapper[4809]: I1205 11:26:34.589289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.589394 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.589467 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:36.589450144 +0000 UTC m=+1091.980426702 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "webhook-server-cert" not found Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.589472 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:26:34 crc kubenswrapper[4809]: E1205 11:26:34.589522 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:36.589508985 +0000 UTC m=+1091.980485543 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "metrics-server-cert" not found Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.061549 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" event={"ID":"2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15","Type":"ContainerStarted","Data":"cf353410b76e24744ff56f0644959d19ea4219b634556ef9ed2f3ddd3776c0a6"} Dec 05 11:26:35 crc kubenswrapper[4809]: E1205 11:26:35.064428 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" podUID="2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15" Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.065370 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" event={"ID":"094859bf-906c-4efa-b894-29db3182659b","Type":"ContainerStarted","Data":"5eeed5aacb2e6f52eda615ecbbf3ca5082fff0086982ff5da041c8ffa49a3fdc"} Dec 05 11:26:35 crc kubenswrapper[4809]: E1205 11:26:35.067432 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" podUID="094859bf-906c-4efa-b894-29db3182659b" Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.068340 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" event={"ID":"990aee67-4ef0-4a34-b228-d511085caf18","Type":"ContainerStarted","Data":"3ccee44d1ef63bb60706d55545fdc17163c6b1aa19d7661b5af57652008fff76"} Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.071560 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" event={"ID":"a738af9f-fa2b-44ea-816a-135b351ef9d2","Type":"ContainerStarted","Data":"db7b59dbcafd61e3bef2f14fd3ada4e54834c123a3c4994feb712e0561dd17a3"} Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.100944 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" event={"ID":"ab4b415e-554a-4e1b-a805-d628772bc6bc","Type":"ContainerStarted","Data":"9f5262c545b79adf6a9bf904dc183f1168566b28830ead018756575aca7902e5"} Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.114363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" event={"ID":"ac756b27-3bb1-4b8d-925c-15b41db189c8","Type":"ContainerStarted","Data":"0cb5714a39aa180081b2adc05e01ea08eda047116fd9308684c584768e04c9d4"} Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.144739 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" event={"ID":"8728b108-ea52-4024-bbba-d1985b5d9795","Type":"ContainerStarted","Data":"040d369f5842deec0379d336f5570dd0f146d41af771d5e3b0a120ef2166763e"} Dec 05 11:26:35 crc kubenswrapper[4809]: E1205 11:26:35.167158 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" podUID="8728b108-ea52-4024-bbba-d1985b5d9795" Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.174143 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" event={"ID":"8bc5d720-4c3a-487b-a58d-4dc4c5760d6e","Type":"ContainerStarted","Data":"313269c0552077be4bc225ddcc08eebd7e2d0f7affc79def807bffd716e2c617"} Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.191857 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" event={"ID":"f586c1d3-4a51-48ff-be7f-d229d2770c10","Type":"ContainerStarted","Data":"2b2c22058fd3e2ab35a214f4da0da7bccd85d94b59190b52dc9384390a53be60"} Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.195289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" event={"ID":"e664f482-5587-4ad2-a35e-0fe439d6de39","Type":"ContainerStarted","Data":"4456ddb1099bc38360eda582c3681062d057223096dfaa379ca8a142288610f4"} Dec 05 11:26:35 crc kubenswrapper[4809]: E1205 11:26:35.204702 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" podUID="e664f482-5587-4ad2-a35e-0fe439d6de39" Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.206190 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" event={"ID":"adcbada3-1e83-4051-a036-f65cce0eb095","Type":"ContainerStarted","Data":"c58af4a663ef9f0ffe234ef0cfef6e2fa0632a1df1fde5a7a79731c24ba2efad"} Dec 05 11:26:35 crc kubenswrapper[4809]: E1205 11:26:35.217901 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" podUID="adcbada3-1e83-4051-a036-f65cce0eb095" Dec 05 11:26:35 crc kubenswrapper[4809]: I1205 11:26:35.818389 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:35 crc kubenswrapper[4809]: E1205 11:26:35.818588 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:35 crc kubenswrapper[4809]: E1205 11:26:35.818655 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert podName:6167f262-8efb-43c5-b065-4a2de04a183f nodeName:}" failed. No retries permitted until 2025-12-05 11:26:39.8186237 +0000 UTC m=+1095.209600258 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert") pod "infra-operator-controller-manager-57548d458d-k8cxw" (UID: "6167f262-8efb-43c5-b065-4a2de04a183f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:36 crc kubenswrapper[4809]: I1205 11:26:36.132569 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.132781 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.133094 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert podName:7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:40.133078042 +0000 UTC m=+1095.524054600 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" (UID: "7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.226961 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" podUID="094859bf-906c-4efa-b894-29db3182659b" Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.227129 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" podUID="8728b108-ea52-4024-bbba-d1985b5d9795" Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.227541 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" podUID="2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15" Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.232961 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" podUID="e664f482-5587-4ad2-a35e-0fe439d6de39" Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.233134 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" podUID="adcbada3-1e83-4051-a036-f65cce0eb095" Dec 05 11:26:36 crc kubenswrapper[4809]: I1205 11:26:36.646911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:36 crc kubenswrapper[4809]: I1205 11:26:36.646963 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.647375 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.647423 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:40.647409743 +0000 UTC m=+1096.038386301 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "metrics-server-cert" not found Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.647461 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:26:36 crc kubenswrapper[4809]: E1205 11:26:36.647478 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:40.647472225 +0000 UTC m=+1096.038448783 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "webhook-server-cert" not found Dec 05 11:26:39 crc kubenswrapper[4809]: I1205 11:26:39.892158 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:39 crc kubenswrapper[4809]: E1205 11:26:39.892358 4809 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:39 crc kubenswrapper[4809]: E1205 11:26:39.893713 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert podName:6167f262-8efb-43c5-b065-4a2de04a183f nodeName:}" failed. No retries permitted until 2025-12-05 11:26:47.893675867 +0000 UTC m=+1103.284652435 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert") pod "infra-operator-controller-manager-57548d458d-k8cxw" (UID: "6167f262-8efb-43c5-b065-4a2de04a183f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:26:40 crc kubenswrapper[4809]: I1205 11:26:40.197514 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:40 crc kubenswrapper[4809]: E1205 11:26:40.197702 4809 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:40 crc kubenswrapper[4809]: E1205 11:26:40.197782 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert podName:7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:48.197765579 +0000 UTC m=+1103.588742137 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" (UID: "7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:26:40 crc kubenswrapper[4809]: I1205 11:26:40.705030 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:40 crc kubenswrapper[4809]: I1205 11:26:40.705204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:40 crc kubenswrapper[4809]: E1205 11:26:40.705340 4809 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:26:40 crc kubenswrapper[4809]: E1205 11:26:40.705395 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:48.705378467 +0000 UTC m=+1104.096355025 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "webhook-server-cert" not found Dec 05 11:26:40 crc kubenswrapper[4809]: E1205 11:26:40.705588 4809 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:26:40 crc kubenswrapper[4809]: E1205 11:26:40.705745 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs podName:69cacfb4-8a61-4bd4-849f-94fc24d85453 nodeName:}" failed. No retries permitted until 2025-12-05 11:26:48.705705766 +0000 UTC m=+1104.096682324 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs") pod "openstack-operator-controller-manager-64b69b8785-6nxr8" (UID: "69cacfb4-8a61-4bd4-849f-94fc24d85453") : secret "metrics-server-cert" not found Dec 05 11:26:45 crc kubenswrapper[4809]: E1205 11:26:45.732783 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 05 11:26:45 crc kubenswrapper[4809]: E1205 11:26:45.733517 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w7c4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-vf5hg_openstack-operators(a738af9f-fa2b-44ea-816a-135b351ef9d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:26:46 crc kubenswrapper[4809]: E1205 11:26:46.737448 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 05 11:26:46 crc kubenswrapper[4809]: E1205 11:26:46.737817 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s4j7p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-dmgm2_openstack-operators(990aee67-4ef0-4a34-b228-d511085caf18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:26:47 crc kubenswrapper[4809]: E1205 11:26:47.227797 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 05 11:26:47 crc kubenswrapper[4809]: E1205 11:26:47.228044 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fjz2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-9m782_openstack-operators(ac756b27-3bb1-4b8d-925c-15b41db189c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:26:47 crc kubenswrapper[4809]: E1205 11:26:47.783220 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 05 11:26:47 crc kubenswrapper[4809]: E1205 11:26:47.783473 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hslbn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-ctckv_openstack-operators(f586c1d3-4a51-48ff-be7f-d229d2770c10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:26:47 crc kubenswrapper[4809]: I1205 11:26:47.914306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:47 crc kubenswrapper[4809]: I1205 11:26:47.924159 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6167f262-8efb-43c5-b065-4a2de04a183f-cert\") pod \"infra-operator-controller-manager-57548d458d-k8cxw\" (UID: \"6167f262-8efb-43c5-b065-4a2de04a183f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:47 crc kubenswrapper[4809]: I1205 11:26:47.926920 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.218579 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.222130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc\" (UID: \"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.236038 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.725156 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.725258 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.730781 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-metrics-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.730874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69cacfb4-8a61-4bd4-849f-94fc24d85453-webhook-certs\") pod \"openstack-operator-controller-manager-64b69b8785-6nxr8\" (UID: \"69cacfb4-8a61-4bd4-849f-94fc24d85453\") " pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:48 crc kubenswrapper[4809]: I1205 11:26:48.754321 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:49 crc kubenswrapper[4809]: E1205 11:26:49.430015 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 05 11:26:49 crc kubenswrapper[4809]: E1205 11:26:49.430581 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rf62k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-w8rhm_openstack-operators(8bc5d720-4c3a-487b-a58d-4dc4c5760d6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:26:49 crc kubenswrapper[4809]: E1205 11:26:49.432371 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" podUID="8bc5d720-4c3a-487b-a58d-4dc4c5760d6e" Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.028199 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc"] Dec 05 11:26:50 crc kubenswrapper[4809]: W1205 11:26:50.030761 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c5cc8f1_39d1_4a3a_bc1d_16e3d1175b83.slice/crio-3b00ea39cdc70acd937433870e298b0bdf895d7667cba7e8dd4f5910f51a0d01 WatchSource:0}: Error finding container 3b00ea39cdc70acd937433870e298b0bdf895d7667cba7e8dd4f5910f51a0d01: Status 404 returned error can't find the container with id 3b00ea39cdc70acd937433870e298b0bdf895d7667cba7e8dd4f5910f51a0d01 Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.106729 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j7868,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-plqqf_openstack-operators(88757511-54a6-45b0-8bd8-90dddf1cae9f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.108426 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" podUID="88757511-54a6-45b0-8bd8-90dddf1cae9f" Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.108472 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-slp92,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-9q4zp_openstack-operators(9eec44c6-6e4d-4945-afef-ed95846e7b50): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.109874 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" podUID="9eec44c6-6e4d-4945-afef-ed95846e7b50" Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.144966 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw"] Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.157355 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crs6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-k8cxw_openstack-operators(6167f262-8efb-43c5-b065-4a2de04a183f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.157822 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8"] Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.163058 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-crs6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-k8cxw_openstack-operators(6167f262-8efb-43c5-b065-4a2de04a183f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.164880 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" podUID="6167f262-8efb-43c5-b065-4a2de04a183f" Dec 05 11:26:50 crc kubenswrapper[4809]: W1205 11:26:50.165787 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69cacfb4_8a61_4bd4_849f_94fc24d85453.slice/crio-f1c16c607e6862db9ed0b4ee1764494966e9d723692f905e596e835041ba12a5 WatchSource:0}: Error finding container f1c16c607e6862db9ed0b4ee1764494966e9d723692f905e596e835041ba12a5: Status 404 returned error can't find the container with id f1c16c607e6862db9ed0b4ee1764494966e9d723692f905e596e835041ba12a5 Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.313263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" event={"ID":"f14d52a2-20e5-429a-8285-82ff53430a63","Type":"ContainerStarted","Data":"b79a135acc0617124f56aec8d8e674556a91337c3fa7069ce72cb05885b6ec97"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.315885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" event={"ID":"6167f262-8efb-43c5-b065-4a2de04a183f","Type":"ContainerStarted","Data":"e8eb533b6061d324d729d8a31a7eb5ed964a619ef53fcc103515484847a56e9c"} Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.319613 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" podUID="6167f262-8efb-43c5-b065-4a2de04a183f" Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.322278 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" event={"ID":"d47fc62a-29c5-4ebc-a3b4-45d6ce997907","Type":"ContainerStarted","Data":"8e0cea5f94e768f61f2e7ee1401d1be9174705bfb7815ea6df7f92b217d2bb74"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.331380 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" event={"ID":"69cacfb4-8a61-4bd4-849f-94fc24d85453","Type":"ContainerStarted","Data":"f1c16c607e6862db9ed0b4ee1764494966e9d723692f905e596e835041ba12a5"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.333202 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" event={"ID":"adae4108-5380-4358-b8bd-44ee6bb554af","Type":"ContainerStarted","Data":"f6495d77a452d2feed39c3e6410043869580129f6ce175aefd64aad253d69478"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.353375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" event={"ID":"ab4b415e-554a-4e1b-a805-d628772bc6bc","Type":"ContainerStarted","Data":"75977068bb6d9e4bd833b1b0b6ad4928ddbfe7086a3f9f958539e52a3e8f93f4"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.363510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" event={"ID":"88757511-54a6-45b0-8bd8-90dddf1cae9f","Type":"ContainerStarted","Data":"389ce4b2b17d4aa1dbbd6ec55557cebfb34c3b24dd539bafabb30d643036dd95"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.365589 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.368843 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" podUID="88757511-54a6-45b0-8bd8-90dddf1cae9f" Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.397533 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" event={"ID":"9eec44c6-6e4d-4945-afef-ed95846e7b50","Type":"ContainerStarted","Data":"770baf6a5443358cdd029e63fd82a5b4d70bbacc8ac815e0349779a7e88a7fda"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.398303 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.407834 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" podUID="9eec44c6-6e4d-4945-afef-ed95846e7b50" Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.410227 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" event={"ID":"84c8e973-2683-4917-bea9-a7f44eadf397","Type":"ContainerStarted","Data":"160e0a27c0eab7d022c517d589e76c52b5bf344f98623b9085bfc834e406bfb6"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.435666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" event={"ID":"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83","Type":"ContainerStarted","Data":"3b00ea39cdc70acd937433870e298b0bdf895d7667cba7e8dd4f5910f51a0d01"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.462269 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" event={"ID":"06bcaa51-2671-458b-9cd0-a4c73455b02a","Type":"ContainerStarted","Data":"bc4e126e6a9d6b596719777387e97f66ec55d443ea18b9903b8457cdae76c796"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.472789 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" event={"ID":"53e22095-69df-44e2-aaea-7a6fbbbfce6c","Type":"ContainerStarted","Data":"37169cfe9ae3b55c42bc5cdd157b371b69b257ac83fd740364a69eb27c319a1a"} Dec 05 11:26:50 crc kubenswrapper[4809]: I1205 11:26:50.483695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" event={"ID":"ad1e2739-6610-4b87-aec7-89ca78575409","Type":"ContainerStarted","Data":"d248fd87c759bcb71435cd5aca3599fb2e3b75b9f29db568a49fee3766e9d9eb"} Dec 05 11:26:50 crc kubenswrapper[4809]: E1205 11:26:50.484782 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" podUID="8bc5d720-4c3a-487b-a58d-4dc4c5760d6e" Dec 05 11:26:51 crc kubenswrapper[4809]: I1205 11:26:51.493660 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" event={"ID":"69cacfb4-8a61-4bd4-849f-94fc24d85453","Type":"ContainerStarted","Data":"11bfe9e38c90aa3beaf25df7a99f64a7340bb9435c49a900b6ce3a89c7062ee2"} Dec 05 11:26:51 crc kubenswrapper[4809]: I1205 11:26:51.494570 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:26:51 crc kubenswrapper[4809]: E1205 11:26:51.496051 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" podUID="9eec44c6-6e4d-4945-afef-ed95846e7b50" Dec 05 11:26:51 crc kubenswrapper[4809]: E1205 11:26:51.496374 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" podUID="88757511-54a6-45b0-8bd8-90dddf1cae9f" Dec 05 11:26:51 crc kubenswrapper[4809]: E1205 11:26:51.499839 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" podUID="6167f262-8efb-43c5-b065-4a2de04a183f" Dec 05 11:26:51 crc kubenswrapper[4809]: I1205 11:26:51.549077 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" podStartSLOduration=19.549060073 podStartE2EDuration="19.549060073s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:26:51.547548372 +0000 UTC m=+1106.938524950" watchObservedRunningTime="2025-12-05 11:26:51.549060073 +0000 UTC m=+1106.940036631" Dec 05 11:26:58 crc kubenswrapper[4809]: I1205 11:26:58.771335 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-64b69b8785-6nxr8" Dec 05 11:27:01 crc kubenswrapper[4809]: I1205 11:27:01.876750 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" Dec 05 11:27:02 crc kubenswrapper[4809]: I1205 11:27:02.226037 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" Dec 05 11:27:13 crc kubenswrapper[4809]: E1205 11:27:13.044004 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 05 11:27:13 crc kubenswrapper[4809]: E1205 11:27:13.045304 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8nl4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-khmw4_openstack-operators(adcbada3-1e83-4051-a036-f65cce0eb095): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:14 crc kubenswrapper[4809]: I1205 11:27:14.047034 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:27:14 crc kubenswrapper[4809]: I1205 11:27:14.047325 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:27:15 crc kubenswrapper[4809]: E1205 11:27:15.714817 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:15 crc kubenswrapper[4809]: E1205 11:27:15.714970 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s4j7p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-dmgm2_openstack-operators(990aee67-4ef0-4a34-b228-d511085caf18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:15 crc kubenswrapper[4809]: E1205 11:27:15.716143 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" podUID="990aee67-4ef0-4a34-b228-d511085caf18" Dec 05 11:27:15 crc kubenswrapper[4809]: E1205 11:27:15.800805 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:15 crc kubenswrapper[4809]: E1205 11:27:15.801255 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fjz2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-9m782_openstack-operators(ac756b27-3bb1-4b8d-925c-15b41db189c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:15 crc kubenswrapper[4809]: E1205 11:27:15.802452 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" podUID="ac756b27-3bb1-4b8d-925c-15b41db189c8" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.143738 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.144223 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2bn5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-cjt5m_openstack-operators(adae4108-5380-4358-b8bd-44ee6bb554af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.145426 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" podUID="adae4108-5380-4358-b8bd-44ee6bb554af" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.157388 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.157617 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qw2nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-5sz6t_openstack-operators(8728b108-ea52-4024-bbba-d1985b5d9795): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.171117 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.171314 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hslbn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-ctckv_openstack-operators(f586c1d3-4a51-48ff-be7f-d229d2770c10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.172516 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" podUID="f586c1d3-4a51-48ff-be7f-d229d2770c10" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.210342 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.210521 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t8bqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-mh2r6_openstack-operators(53e22095-69df-44e2-aaea-7a6fbbbfce6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.212782 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" podUID="53e22095-69df-44e2-aaea-7a6fbbbfce6c" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.685386 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" podUID="53e22095-69df-44e2-aaea-7a6fbbbfce6c" Dec 05 11:27:16 crc kubenswrapper[4809]: E1205 11:27:16.685501 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" podUID="adae4108-5380-4358-b8bd-44ee6bb554af" Dec 05 11:27:16 crc kubenswrapper[4809]: I1205 11:27:16.687420 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" Dec 05 11:27:16 crc kubenswrapper[4809]: I1205 11:27:16.690233 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.028865 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.029032 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mqdjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-57zmz_openstack-operators(ab4b415e-554a-4e1b-a805-d628772bc6bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.030785 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" podUID="ab4b415e-554a-4e1b-a805-d628772bc6bc" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.046814 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.047114 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x4lww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-c27fb_openstack-operators(094859bf-906c-4efa-b894-29db3182659b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.063365 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.063490 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2cwbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-jgxpj_openstack-operators(06bcaa51-2671-458b-9cd0-a4c73455b02a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.064706 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" podUID="06bcaa51-2671-458b-9cd0-a4c73455b02a" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.092405 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.092550 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p4ccd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-nmg6x_openstack-operators(ad1e2739-6610-4b87-aec7-89ca78575409): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.093101 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.093298 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7rpbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-hx8w2_openstack-operators(d47fc62a-29c5-4ebc-a3b4-45d6ce997907): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.094437 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" podUID="ad1e2739-6610-4b87-aec7-89ca78575409" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.094451 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" podUID="d47fc62a-29c5-4ebc-a3b4-45d6ce997907" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.546076 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" podUID="adcbada3-1e83-4051-a036-f65cce0eb095" Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.560031 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" podUID="a738af9f-fa2b-44ea-816a-135b351ef9d2" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.706074 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" event={"ID":"2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15","Type":"ContainerStarted","Data":"5a944e833f29742cf42f0428f3013cac48e0273cb3ce877438ccafee662001c6"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.707237 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.720218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" event={"ID":"88757511-54a6-45b0-8bd8-90dddf1cae9f","Type":"ContainerStarted","Data":"eb918deb3499bb4b23a0ae1aa58b67dbf8781feaaef1e1981a499b0ce1d21e6c"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.730646 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" podStartSLOduration=3.722865384 podStartE2EDuration="46.73061754s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.071171556 +0000 UTC m=+1089.462148114" lastFinishedPulling="2025-12-05 11:27:17.078923712 +0000 UTC m=+1132.469900270" observedRunningTime="2025-12-05 11:27:17.72620593 +0000 UTC m=+1133.117182488" watchObservedRunningTime="2025-12-05 11:27:17.73061754 +0000 UTC m=+1133.121594098" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.735524 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" event={"ID":"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83","Type":"ContainerStarted","Data":"a56acbb05df628b88c5274718bd6585a09f81cecef1e839212cd1d3192bef77a"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.736049 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.753726 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-plqqf" podStartSLOduration=2.862666744 podStartE2EDuration="46.753704008s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.342909305 +0000 UTC m=+1088.733885863" lastFinishedPulling="2025-12-05 11:27:17.233946519 +0000 UTC m=+1132.624923127" observedRunningTime="2025-12-05 11:27:17.7519573 +0000 UTC m=+1133.142933868" watchObservedRunningTime="2025-12-05 11:27:17.753704008 +0000 UTC m=+1133.144680566" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.758551 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" event={"ID":"8bc5d720-4c3a-487b-a58d-4dc4c5760d6e","Type":"ContainerStarted","Data":"19e15e40f04dd88eac35cd20543d0201489c8935f4ec0b4111a3b01766bdd7b5"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.792362 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" podStartSLOduration=18.727429365 podStartE2EDuration="45.792343119s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:50.033364035 +0000 UTC m=+1105.424340593" lastFinishedPulling="2025-12-05 11:27:17.098277789 +0000 UTC m=+1132.489254347" observedRunningTime="2025-12-05 11:27:17.789172153 +0000 UTC m=+1133.180148711" watchObservedRunningTime="2025-12-05 11:27:17.792343119 +0000 UTC m=+1133.183319677" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.794614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" event={"ID":"a738af9f-fa2b-44ea-816a-135b351ef9d2","Type":"ContainerStarted","Data":"319ff4f14bd55d21f5fdffeea29dad17b1ec3bad5205841c30fac5a3745ee066"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.807001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" event={"ID":"9eec44c6-6e4d-4945-afef-ed95846e7b50","Type":"ContainerStarted","Data":"ce275dbb0da531d659901082a733bef2e89b57af5b99ba3988b18de6fb138380"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.830142 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-w8rhm" podStartSLOduration=2.723337456 podStartE2EDuration="45.830117126s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.998677794 +0000 UTC m=+1089.389654352" lastFinishedPulling="2025-12-05 11:27:17.105457464 +0000 UTC m=+1132.496434022" observedRunningTime="2025-12-05 11:27:17.811305605 +0000 UTC m=+1133.202282173" watchObservedRunningTime="2025-12-05 11:27:17.830117126 +0000 UTC m=+1133.221093684" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.852872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" event={"ID":"84c8e973-2683-4917-bea9-a7f44eadf397","Type":"ContainerStarted","Data":"fbf5af60f02b1e82fb001eb9e0306886905e79dc8202ccbf1570e68d674ea4bc"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.854850 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.859164 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.861995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" event={"ID":"adcbada3-1e83-4051-a036-f65cce0eb095","Type":"ContainerStarted","Data":"b27a814df7d1529d7541fb704742f4d6a6e21064c3e1b9e0d9e2c65e6788519d"} Dec 05 11:27:17 crc kubenswrapper[4809]: E1205 11:27:17.866817 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" podUID="adcbada3-1e83-4051-a036-f65cce0eb095" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.867861 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-9q4zp" podStartSLOduration=3.453902567 podStartE2EDuration="46.867843183s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.750343308 +0000 UTC m=+1089.141319866" lastFinishedPulling="2025-12-05 11:27:17.164283914 +0000 UTC m=+1132.555260482" observedRunningTime="2025-12-05 11:27:17.836990423 +0000 UTC m=+1133.227966991" watchObservedRunningTime="2025-12-05 11:27:17.867843183 +0000 UTC m=+1133.258819741" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.885895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" event={"ID":"f14d52a2-20e5-429a-8285-82ff53430a63","Type":"ContainerStarted","Data":"689080a8a2d4981982963bdbd0ae442328f849cc86fa1489ab565efa4631b0e5"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.886538 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.902959 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-57hcg" podStartSLOduration=3.129926673 podStartE2EDuration="46.902940257s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.33132891 +0000 UTC m=+1088.722305468" lastFinishedPulling="2025-12-05 11:27:17.104342494 +0000 UTC m=+1132.495319052" observedRunningTime="2025-12-05 11:27:17.886089099 +0000 UTC m=+1133.277065657" watchObservedRunningTime="2025-12-05 11:27:17.902940257 +0000 UTC m=+1133.293916815" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.922336 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.923978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" event={"ID":"e664f482-5587-4ad2-a35e-0fe439d6de39","Type":"ContainerStarted","Data":"c75f1eba83da06a66f3ebe7d965009e9a0aa6dd9d9ebaab92afb81d2f89a4f55"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.924014 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" event={"ID":"e664f482-5587-4ad2-a35e-0fe439d6de39","Type":"ContainerStarted","Data":"1ded845937920b9173c46ed1f76f9c531020fbc7a79647b34ef3c7f6a5a8a8fe"} Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.924027 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.925520 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.925559 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.925577 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.925590 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.944867 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-cc22d" podStartSLOduration=2.830777026 podStartE2EDuration="46.944844127s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.035787441 +0000 UTC m=+1088.426763999" lastFinishedPulling="2025-12-05 11:27:17.149854542 +0000 UTC m=+1132.540831100" observedRunningTime="2025-12-05 11:27:17.922835819 +0000 UTC m=+1133.313812377" watchObservedRunningTime="2025-12-05 11:27:17.944844127 +0000 UTC m=+1133.335820685" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.955707 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.957650 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" Dec 05 11:27:17 crc kubenswrapper[4809]: I1205 11:27:17.963907 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.055886 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" Dec 05 11:27:18 crc kubenswrapper[4809]: E1205 11:27:18.059999 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" podUID="8728b108-ea52-4024-bbba-d1985b5d9795" Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.091472 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" podStartSLOduration=4.049164138 podStartE2EDuration="47.091454345s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.064210356 +0000 UTC m=+1089.455186914" lastFinishedPulling="2025-12-05 11:27:17.106500563 +0000 UTC m=+1132.497477121" observedRunningTime="2025-12-05 11:27:18.09015037 +0000 UTC m=+1133.481126928" watchObservedRunningTime="2025-12-05 11:27:18.091454345 +0000 UTC m=+1133.482430903" Dec 05 11:27:18 crc kubenswrapper[4809]: E1205 11:27:18.659054 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" podUID="094859bf-906c-4efa-b894-29db3182659b" Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.936236 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" event={"ID":"53e22095-69df-44e2-aaea-7a6fbbbfce6c","Type":"ContainerStarted","Data":"744a9a9b0f733cb4732f6921b5a9b5dcb2694468081aa5e8ab90f81fb31f30a5"} Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.938082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" event={"ID":"094859bf-906c-4efa-b894-29db3182659b","Type":"ContainerStarted","Data":"eb143fb87874cd7f21aa568cafb567fe423537a2f42046830fcf8198ed13e836"} Dec 05 11:27:18 crc kubenswrapper[4809]: E1205 11:27:18.940501 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" podUID="094859bf-906c-4efa-b894-29db3182659b" Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.941049 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" event={"ID":"990aee67-4ef0-4a34-b228-d511085caf18","Type":"ContainerStarted","Data":"105cf8397d759b91402d7e20550f97ebfbda0aed2a94e937e447bcd592671572"} Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.947077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" event={"ID":"a738af9f-fa2b-44ea-816a-135b351ef9d2","Type":"ContainerStarted","Data":"a9e6d5027e940eb72006abb41d8ada1e253e971375f72fa1660366af81c3324c"} Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.947736 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.953206 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" event={"ID":"ad1e2739-6610-4b87-aec7-89ca78575409","Type":"ContainerStarted","Data":"553d5f4e14c819c0b229614a52464879f8b929d9d6a9c40ca7130b33f1be1318"} Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.960449 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" event={"ID":"06bcaa51-2671-458b-9cd0-a4c73455b02a","Type":"ContainerStarted","Data":"e9272b9d864b0aa9555a217315b7e0b2360a00633b736090a02cc62000818bad"} Dec 05 11:27:18 crc kubenswrapper[4809]: I1205 11:27:18.973071 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" event={"ID":"8728b108-ea52-4024-bbba-d1985b5d9795","Type":"ContainerStarted","Data":"cf62a858be381ef47803cec28dd5728e0abe3d8ac5c355fe1eeb011eee720a7f"} Dec 05 11:27:18 crc kubenswrapper[4809]: E1205 11:27:18.982962 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" podUID="8728b108-ea52-4024-bbba-d1985b5d9795" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.006704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" event={"ID":"2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15","Type":"ContainerStarted","Data":"2b8f4038fc9608275f5f6f488a5d6f68b2ab7effc6d5166c731f85fbba36dc53"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.014037 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mh2r6" podStartSLOduration=31.310939099 podStartE2EDuration="47.01401511s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.786740168 +0000 UTC m=+1089.177716726" lastFinishedPulling="2025-12-05 11:26:49.489816179 +0000 UTC m=+1104.880792737" observedRunningTime="2025-12-05 11:27:18.974246528 +0000 UTC m=+1134.365223086" watchObservedRunningTime="2025-12-05 11:27:19.01401511 +0000 UTC m=+1134.404991668" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.034075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" event={"ID":"d47fc62a-29c5-4ebc-a3b4-45d6ce997907","Type":"ContainerStarted","Data":"617698363aff660c81c2e26fc54947dd5f5f13f702fda8aeb3e4c39e081c73d9"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.039154 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" podStartSLOduration=2.576547322 podStartE2EDuration="47.039140173s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.019200232 +0000 UTC m=+1089.410176790" lastFinishedPulling="2025-12-05 11:27:18.481793083 +0000 UTC m=+1133.872769641" observedRunningTime="2025-12-05 11:27:19.016349543 +0000 UTC m=+1134.407326101" watchObservedRunningTime="2025-12-05 11:27:19.039140173 +0000 UTC m=+1134.430116731" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.044015 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-jgxpj" podStartSLOduration=32.289217628 podStartE2EDuration="48.044003435s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.746319989 +0000 UTC m=+1089.137296547" lastFinishedPulling="2025-12-05 11:26:49.501105796 +0000 UTC m=+1104.892082354" observedRunningTime="2025-12-05 11:27:19.038557347 +0000 UTC m=+1134.429533905" watchObservedRunningTime="2025-12-05 11:27:19.044003435 +0000 UTC m=+1134.434979993" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.064469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" event={"ID":"7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83","Type":"ContainerStarted","Data":"fe87700f6233847d35538b19dc866280a03653fa3c12ff5eb88afb79c52f5f8a"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.073551 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-nmg6x" podStartSLOduration=32.358133013 podStartE2EDuration="48.073527999s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.751431808 +0000 UTC m=+1089.142408366" lastFinishedPulling="2025-12-05 11:26:49.466826794 +0000 UTC m=+1104.857803352" observedRunningTime="2025-12-05 11:27:19.068586524 +0000 UTC m=+1134.459563092" watchObservedRunningTime="2025-12-05 11:27:19.073527999 +0000 UTC m=+1134.464504557" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.080222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" event={"ID":"f586c1d3-4a51-48ff-be7f-d229d2770c10","Type":"ContainerStarted","Data":"2cc05c89acdb79e3f287336a008bc600953f42c954eb210302f08aaf3608388d"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.080495 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" event={"ID":"f586c1d3-4a51-48ff-be7f-d229d2770c10","Type":"ContainerStarted","Data":"ce2cb75ed283cb4a79411f8424301189e0aa7e9bcb90ad82d16057fe2ea57444"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.081346 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.101996 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" event={"ID":"ab4b415e-554a-4e1b-a805-d628772bc6bc","Type":"ContainerStarted","Data":"657179ad0b2afc62980fe0840252c38d5e413dfd838fd17ff927e54a95a6bb3f"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.118580 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" event={"ID":"ac756b27-3bb1-4b8d-925c-15b41db189c8","Type":"ContainerStarted","Data":"8cb57906fd725ef31a04210f26d8bcf26b40eb6d77479c260e516ff1a13039c3"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.118644 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" event={"ID":"ac756b27-3bb1-4b8d-925c-15b41db189c8","Type":"ContainerStarted","Data":"314b773800d6bc82549d52a2ed837de85e6aa008668589db746310ea71834082"} Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.119077 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.150470 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" podStartSLOduration=3.602942041 podStartE2EDuration="47.150452721s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.185824054 +0000 UTC m=+1089.576800612" lastFinishedPulling="2025-12-05 11:27:17.733334734 +0000 UTC m=+1133.124311292" observedRunningTime="2025-12-05 11:27:19.145007343 +0000 UTC m=+1134.535983901" watchObservedRunningTime="2025-12-05 11:27:19.150452721 +0000 UTC m=+1134.541429279" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.187612 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-hx8w2" podStartSLOduration=31.469175503 podStartE2EDuration="47.187598251s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.753581166 +0000 UTC m=+1089.144557724" lastFinishedPulling="2025-12-05 11:26:49.472003914 +0000 UTC m=+1104.862980472" observedRunningTime="2025-12-05 11:27:19.181398893 +0000 UTC m=+1134.572375441" watchObservedRunningTime="2025-12-05 11:27:19.187598251 +0000 UTC m=+1134.578574809" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.235194 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-57zmz" podStartSLOduration=32.75798583 podStartE2EDuration="48.235179436s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.019581142 +0000 UTC m=+1089.410557700" lastFinishedPulling="2025-12-05 11:26:49.496774748 +0000 UTC m=+1104.887751306" observedRunningTime="2025-12-05 11:27:19.231728502 +0000 UTC m=+1134.622705060" watchObservedRunningTime="2025-12-05 11:27:19.235179436 +0000 UTC m=+1134.626155994" Dec 05 11:27:19 crc kubenswrapper[4809]: I1205 11:27:19.253233 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" podStartSLOduration=3.545365995 podStartE2EDuration="47.253215856s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.019281134 +0000 UTC m=+1089.410257692" lastFinishedPulling="2025-12-05 11:27:17.727130995 +0000 UTC m=+1133.118107553" observedRunningTime="2025-12-05 11:27:19.252364063 +0000 UTC m=+1134.643340611" watchObservedRunningTime="2025-12-05 11:27:19.253215856 +0000 UTC m=+1134.644192414" Dec 05 11:27:20 crc kubenswrapper[4809]: I1205 11:27:20.130691 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" event={"ID":"990aee67-4ef0-4a34-b228-d511085caf18","Type":"ContainerStarted","Data":"72c58e465e7e4fef909c6430002a9fa28c1828b1ff561313beced434d0829e46"} Dec 05 11:27:20 crc kubenswrapper[4809]: I1205 11:27:20.132601 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" Dec 05 11:27:20 crc kubenswrapper[4809]: I1205 11:27:20.154532 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" podStartSLOduration=4.486403172 podStartE2EDuration="49.154513534s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.037836269 +0000 UTC m=+1089.428812827" lastFinishedPulling="2025-12-05 11:27:18.705946631 +0000 UTC m=+1134.096923189" observedRunningTime="2025-12-05 11:27:20.147466022 +0000 UTC m=+1135.538442580" watchObservedRunningTime="2025-12-05 11:27:20.154513534 +0000 UTC m=+1135.545490092" Dec 05 11:27:21 crc kubenswrapper[4809]: I1205 11:27:21.139564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" event={"ID":"6167f262-8efb-43c5-b065-4a2de04a183f","Type":"ContainerStarted","Data":"83d9114d51c0d0187eec44f7bd4f6a5febc1b50efa8cb65a155f142a37e792de"} Dec 05 11:27:21 crc kubenswrapper[4809]: I1205 11:27:21.139972 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" event={"ID":"6167f262-8efb-43c5-b065-4a2de04a183f","Type":"ContainerStarted","Data":"c57a8e308272a17c287c5fe00ae84bf5850e701730a79ecacf600aaadc4d3477"} Dec 05 11:27:21 crc kubenswrapper[4809]: I1205 11:27:21.140206 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:27:21 crc kubenswrapper[4809]: I1205 11:27:21.161346 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" podStartSLOduration=20.039248176 podStartE2EDuration="50.161324091s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:50.157225373 +0000 UTC m=+1105.548201931" lastFinishedPulling="2025-12-05 11:27:20.279301288 +0000 UTC m=+1135.670277846" observedRunningTime="2025-12-05 11:27:21.15541744 +0000 UTC m=+1136.546394018" watchObservedRunningTime="2025-12-05 11:27:21.161324091 +0000 UTC m=+1136.552300669" Dec 05 11:27:22 crc kubenswrapper[4809]: I1205 11:27:22.161672 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" Dec 05 11:27:22 crc kubenswrapper[4809]: I1205 11:27:22.164167 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" Dec 05 11:27:22 crc kubenswrapper[4809]: I1205 11:27:22.352904 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-jdbkf" Dec 05 11:27:22 crc kubenswrapper[4809]: I1205 11:27:22.437145 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-7z7pv" Dec 05 11:27:23 crc kubenswrapper[4809]: I1205 11:27:23.052260 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ctckv" Dec 05 11:27:23 crc kubenswrapper[4809]: I1205 11:27:23.144875 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-vf5hg" Dec 05 11:27:23 crc kubenswrapper[4809]: I1205 11:27:23.158107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" event={"ID":"adae4108-5380-4358-b8bd-44ee6bb554af","Type":"ContainerStarted","Data":"eabb585e863da30d1731c8566cf68d2de8076c762a9027c743385bbda3e89dba"} Dec 05 11:27:23 crc kubenswrapper[4809]: I1205 11:27:23.185362 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-cjt5m" podStartSLOduration=36.038208667 podStartE2EDuration="52.185340627s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:33.324271898 +0000 UTC m=+1088.715248456" lastFinishedPulling="2025-12-05 11:26:49.471403858 +0000 UTC m=+1104.862380416" observedRunningTime="2025-12-05 11:27:23.18286733 +0000 UTC m=+1138.573843908" watchObservedRunningTime="2025-12-05 11:27:23.185340627 +0000 UTC m=+1138.576317185" Dec 05 11:27:27 crc kubenswrapper[4809]: I1205 11:27:27.933453 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-k8cxw" Dec 05 11:27:28 crc kubenswrapper[4809]: I1205 11:27:28.246891 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc" Dec 05 11:27:28 crc kubenswrapper[4809]: E1205 11:27:28.877112 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" podUID="adcbada3-1e83-4051-a036-f65cce0eb095" Dec 05 11:27:29 crc kubenswrapper[4809]: E1205 11:27:29.881213 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" podUID="094859bf-906c-4efa-b894-29db3182659b" Dec 05 11:27:30 crc kubenswrapper[4809]: E1205 11:27:30.875365 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" podUID="8728b108-ea52-4024-bbba-d1985b5d9795" Dec 05 11:27:32 crc kubenswrapper[4809]: I1205 11:27:32.421331 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-dmgm2" Dec 05 11:27:32 crc kubenswrapper[4809]: I1205 11:27:32.863807 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-9m782" Dec 05 11:27:43 crc kubenswrapper[4809]: I1205 11:27:43.873918 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:27:44 crc kubenswrapper[4809]: I1205 11:27:44.046486 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:27:44 crc kubenswrapper[4809]: I1205 11:27:44.046553 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:27:44 crc kubenswrapper[4809]: I1205 11:27:44.322522 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" event={"ID":"094859bf-906c-4efa-b894-29db3182659b","Type":"ContainerStarted","Data":"b583a200b60bae149dd0f73e060501fbf6801ace61694e86fb22a27bc4b0f56e"} Dec 05 11:27:44 crc kubenswrapper[4809]: I1205 11:27:44.322736 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" Dec 05 11:27:44 crc kubenswrapper[4809]: I1205 11:27:44.337802 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" podStartSLOduration=4.013213 podStartE2EDuration="1m13.337755048s" podCreationTimestamp="2025-12-05 11:26:31 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.04963841 +0000 UTC m=+1089.440614968" lastFinishedPulling="2025-12-05 11:27:43.374180458 +0000 UTC m=+1158.765157016" observedRunningTime="2025-12-05 11:27:44.336863914 +0000 UTC m=+1159.727840472" watchObservedRunningTime="2025-12-05 11:27:44.337755048 +0000 UTC m=+1159.728731606" Dec 05 11:27:45 crc kubenswrapper[4809]: I1205 11:27:45.331266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" event={"ID":"adcbada3-1e83-4051-a036-f65cce0eb095","Type":"ContainerStarted","Data":"d0c69112227fdb2435381eb229d55e1c1d5c380c7fe42dbfe4c63ac94011f07c"} Dec 05 11:27:45 crc kubenswrapper[4809]: I1205 11:27:45.331771 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" Dec 05 11:27:45 crc kubenswrapper[4809]: I1205 11:27:45.351545 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" podStartSLOduration=2.571461145 podStartE2EDuration="1m13.351527175s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.051885351 +0000 UTC m=+1089.442861909" lastFinishedPulling="2025-12-05 11:27:44.831951381 +0000 UTC m=+1160.222927939" observedRunningTime="2025-12-05 11:27:45.345553322 +0000 UTC m=+1160.736529880" watchObservedRunningTime="2025-12-05 11:27:45.351527175 +0000 UTC m=+1160.742503733" Dec 05 11:27:47 crc kubenswrapper[4809]: I1205 11:27:47.351784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" event={"ID":"8728b108-ea52-4024-bbba-d1985b5d9795","Type":"ContainerStarted","Data":"d5bfab6ae65bcda8ec08efb87900600bc7b860a547fb9b48c9790cad7aa410da"} Dec 05 11:27:47 crc kubenswrapper[4809]: I1205 11:27:47.352344 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" Dec 05 11:27:47 crc kubenswrapper[4809]: I1205 11:27:47.372227 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" podStartSLOduration=3.121117346 podStartE2EDuration="1m15.372210762s" podCreationTimestamp="2025-12-05 11:26:32 +0000 UTC" firstStartedPulling="2025-12-05 11:26:34.053191786 +0000 UTC m=+1089.444168344" lastFinishedPulling="2025-12-05 11:27:46.304285192 +0000 UTC m=+1161.695261760" observedRunningTime="2025-12-05 11:27:47.368494251 +0000 UTC m=+1162.759470819" watchObservedRunningTime="2025-12-05 11:27:47.372210762 +0000 UTC m=+1162.763187320" Dec 05 11:27:52 crc kubenswrapper[4809]: I1205 11:27:52.555590 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-c27fb" Dec 05 11:27:52 crc kubenswrapper[4809]: I1205 11:27:52.841183 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5sz6t" Dec 05 11:27:53 crc kubenswrapper[4809]: I1205 11:27:53.123082 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-khmw4" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.075999 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd7lk"] Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.078012 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.079997 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ppw84" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.080365 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.080564 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.080794 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.100390 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd7lk"] Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.142935 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zjbwc"] Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.144869 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.146464 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.150664 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zjbwc"] Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.208444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lv4g\" (UniqueName: \"kubernetes.io/projected/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-kube-api-access-7lv4g\") pod \"dnsmasq-dns-675f4bcbfc-kd7lk\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.209039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-config\") pod \"dnsmasq-dns-675f4bcbfc-kd7lk\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.310342 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lv4g\" (UniqueName: \"kubernetes.io/projected/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-kube-api-access-7lv4g\") pod \"dnsmasq-dns-675f4bcbfc-kd7lk\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.310468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-config\") pod \"dnsmasq-dns-675f4bcbfc-kd7lk\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.310504 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-config\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.310541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.310584 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdjh8\" (UniqueName: \"kubernetes.io/projected/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-kube-api-access-kdjh8\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.311429 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-config\") pod \"dnsmasq-dns-675f4bcbfc-kd7lk\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.335891 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lv4g\" (UniqueName: \"kubernetes.io/projected/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-kube-api-access-7lv4g\") pod \"dnsmasq-dns-675f4bcbfc-kd7lk\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.413169 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-config\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.413257 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.413297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdjh8\" (UniqueName: \"kubernetes.io/projected/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-kube-api-access-kdjh8\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.413424 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.414245 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.414262 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-config\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.432790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdjh8\" (UniqueName: \"kubernetes.io/projected/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-kube-api-access-kdjh8\") pod \"dnsmasq-dns-78dd6ddcc-zjbwc\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.466243 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.796670 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zjbwc"] Dec 05 11:28:09 crc kubenswrapper[4809]: I1205 11:28:09.936793 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd7lk"] Dec 05 11:28:09 crc kubenswrapper[4809]: W1205 11:28:09.950908 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5b74800_4cb0_4eb3_9d70_7b1935f49ca5.slice/crio-8b0cb67af22987e1fa76145856ab4978d99375841596bcf6df8f172c77267f03 WatchSource:0}: Error finding container 8b0cb67af22987e1fa76145856ab4978d99375841596bcf6df8f172c77267f03: Status 404 returned error can't find the container with id 8b0cb67af22987e1fa76145856ab4978d99375841596bcf6df8f172c77267f03 Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.314976 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd7lk"] Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.350578 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nm8sc"] Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.352038 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.377977 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nm8sc"] Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.430522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.430622 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-config\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.430686 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz4zx\" (UniqueName: \"kubernetes.io/projected/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-kube-api-access-jz4zx\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.532513 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-config\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.532579 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz4zx\" (UniqueName: \"kubernetes.io/projected/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-kube-api-access-jz4zx\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.532622 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.534159 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.534761 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-config\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.548270 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" event={"ID":"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5","Type":"ContainerStarted","Data":"8b0cb67af22987e1fa76145856ab4978d99375841596bcf6df8f172c77267f03"} Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.549391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" event={"ID":"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087","Type":"ContainerStarted","Data":"9c02a4cb0427f5048ff7ccff44d9dce48e7ca3e24d8cf82cbaa759e182c665c6"} Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.558153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz4zx\" (UniqueName: \"kubernetes.io/projected/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-kube-api-access-jz4zx\") pod \"dnsmasq-dns-666b6646f7-nm8sc\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:10 crc kubenswrapper[4809]: I1205 11:28:10.697845 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.126375 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zjbwc"] Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.156520 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-49vg4"] Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.157740 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.196001 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-49vg4"] Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.251706 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.251780 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-config\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.251872 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj9sw\" (UniqueName: \"kubernetes.io/projected/f38368a6-2aa9-454d-bf49-816d959bc81c-kube-api-access-dj9sw\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.353760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.353837 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-config\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.353917 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj9sw\" (UniqueName: \"kubernetes.io/projected/f38368a6-2aa9-454d-bf49-816d959bc81c-kube-api-access-dj9sw\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.354901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.355443 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-config\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.384047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj9sw\" (UniqueName: \"kubernetes.io/projected/f38368a6-2aa9-454d-bf49-816d959bc81c-kube-api-access-dj9sw\") pod \"dnsmasq-dns-57d769cc4f-49vg4\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.499903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.503490 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.506146 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.507911 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sq4zw" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.507924 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.508276 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.508352 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.508451 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.508531 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.508615 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.522014 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nm8sc"] Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.548678 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:28:11 crc kubenswrapper[4809]: W1205 11:28:11.570299 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5a72a5a_f9ed_483c_bc9d_1866d9a29ce2.slice/crio-5ec9778ceb64b25ebb773904ab9a314cce960c967b771d01e6ea99224f300080 WatchSource:0}: Error finding container 5ec9778ceb64b25ebb773904ab9a314cce960c967b771d01e6ea99224f300080: Status 404 returned error can't find the container with id 5ec9778ceb64b25ebb773904ab9a314cce960c967b771d01e6ea99224f300080 Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657833 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e1251531-cb99-418e-aa38-65a793e5a9d0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657950 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.657998 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e1251531-cb99-418e-aa38-65a793e5a9d0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.658025 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.658046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.658072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpxvp\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-kube-api-access-rpxvp\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759313 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759733 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e1251531-cb99-418e-aa38-65a793e5a9d0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759753 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759769 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpxvp\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-kube-api-access-rpxvp\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759836 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759868 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.759919 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e1251531-cb99-418e-aa38-65a793e5a9d0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.760404 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.760708 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.760838 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.761185 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.761334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.763883 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.764026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e1251531-cb99-418e-aa38-65a793e5a9d0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.764863 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.771387 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e1251531-cb99-418e-aa38-65a793e5a9d0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.771483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.782869 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpxvp\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-kube-api-access-rpxvp\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.814833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.848290 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 11:28:11 crc kubenswrapper[4809]: I1205 11:28:11.868663 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-49vg4"] Dec 05 11:28:11 crc kubenswrapper[4809]: W1205 11:28:11.946378 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf38368a6_2aa9_454d_bf49_816d959bc81c.slice/crio-d14b51c5a0cf76b5c728ee1c78e1d4b66ea0b6f7ef968d5d3e3aaa44bcd31811 WatchSource:0}: Error finding container d14b51c5a0cf76b5c728ee1c78e1d4b66ea0b6f7ef968d5d3e3aaa44bcd31811: Status 404 returned error can't find the container with id d14b51c5a0cf76b5c728ee1c78e1d4b66ea0b6f7ef968d5d3e3aaa44bcd31811 Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.344847 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.346847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.348255 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.352705 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.353060 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-f2xpn" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.354572 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.354875 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.355029 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.356240 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.356504 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.435466 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:28:12 crc kubenswrapper[4809]: W1205 11:28:12.446569 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1251531_cb99_418e_aa38_65a793e5a9d0.slice/crio-fd93b120571515d66e8da7967dc33a5338e633e211a80c2962cea3e0498ea404 WatchSource:0}: Error finding container fd93b120571515d66e8da7967dc33a5338e633e211a80c2962cea3e0498ea404: Status 404 returned error can't find the container with id fd93b120571515d66e8da7967dc33a5338e633e211a80c2962cea3e0498ea404 Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.471975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfe522ab-449f-433e-a9bf-b33ed3b42595-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqxrg\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-kube-api-access-sqxrg\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472113 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfe522ab-449f-433e-a9bf-b33ed3b42595-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472150 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472187 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.472330 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.573961 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfe522ab-449f-433e-a9bf-b33ed3b42595-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574025 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574064 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqxrg\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-kube-api-access-sqxrg\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574084 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfe522ab-449f-433e-a9bf-b33ed3b42595-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574157 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574223 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574246 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574269 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574377 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.574946 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.578560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.579121 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.581203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.599993 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" event={"ID":"f38368a6-2aa9-454d-bf49-816d959bc81c","Type":"ContainerStarted","Data":"d14b51c5a0cf76b5c728ee1c78e1d4b66ea0b6f7ef968d5d3e3aaa44bcd31811"} Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.602397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" event={"ID":"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2","Type":"ContainerStarted","Data":"5ec9778ceb64b25ebb773904ab9a314cce960c967b771d01e6ea99224f300080"} Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.581936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.605325 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.607221 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.609747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfe522ab-449f-433e-a9bf-b33ed3b42595-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.612658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.616137 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqxrg\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-kube-api-access-sqxrg\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.622113 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e1251531-cb99-418e-aa38-65a793e5a9d0","Type":"ContainerStarted","Data":"fd93b120571515d66e8da7967dc33a5338e633e211a80c2962cea3e0498ea404"} Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.628490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfe522ab-449f-433e-a9bf-b33ed3b42595-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.635011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:12 crc kubenswrapper[4809]: I1205 11:28:12.718182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.386460 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:28:13 crc kubenswrapper[4809]: W1205 11:28:13.419546 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfe522ab_449f_433e_a9bf_b33ed3b42595.slice/crio-63a74e3b8be4d2e13bf4283513c5308ca14b97ff351ef8e5226ff822788891ac WatchSource:0}: Error finding container 63a74e3b8be4d2e13bf4283513c5308ca14b97ff351ef8e5226ff822788891ac: Status 404 returned error can't find the container with id 63a74e3b8be4d2e13bf4283513c5308ca14b97ff351ef8e5226ff822788891ac Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.630130 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.632652 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.636463 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nj98s" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.637166 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.637477 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.637734 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.651067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfe522ab-449f-433e-a9bf-b33ed3b42595","Type":"ContainerStarted","Data":"63a74e3b8be4d2e13bf4283513c5308ca14b97ff351ef8e5226ff822788891ac"} Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.657041 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.665097 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.814781 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.814834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.814878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tncfv\" (UniqueName: \"kubernetes.io/projected/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kube-api-access-tncfv\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.814911 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.815064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kolla-config\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.815132 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-default\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.815179 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.815329 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.916754 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917373 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917418 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tncfv\" (UniqueName: \"kubernetes.io/projected/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kube-api-access-tncfv\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917489 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kolla-config\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-default\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917885 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.917896 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.918846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kolla-config\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.919092 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-default\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.919672 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.925688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.942134 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.942169 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tncfv\" (UniqueName: \"kubernetes.io/projected/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kube-api-access-tncfv\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:13 crc kubenswrapper[4809]: I1205 11:28:13.994435 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " pod="openstack/openstack-galera-0" Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.055647 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.055718 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.055775 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.056467 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f84b335fda3d018d7628fbb24eef2ea4ed19d76e6c066be3637c91567a56e36e"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.056528 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://f84b335fda3d018d7628fbb24eef2ea4ed19d76e6c066be3637c91567a56e36e" gracePeriod=600 Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.257889 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.670973 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="f84b335fda3d018d7628fbb24eef2ea4ed19d76e6c066be3637c91567a56e36e" exitCode=0 Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.671128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"f84b335fda3d018d7628fbb24eef2ea4ed19d76e6c066be3637c91567a56e36e"} Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.673837 4809 scope.go:117] "RemoveContainer" containerID="bd6b669d62c0db11d4d46d14f9c5c9d47f99d27bb3324e58211f411bd95dcbe7" Dec 05 11:28:14 crc kubenswrapper[4809]: I1205 11:28:14.801712 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:28:14 crc kubenswrapper[4809]: W1205 11:28:14.889662 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b760a19_f9d4_4ea7_a6c7_6610aa0334a9.slice/crio-805f4b013934c0f9ea85187e55df41599e25501b8b982044f1c99d18e7dc62e6 WatchSource:0}: Error finding container 805f4b013934c0f9ea85187e55df41599e25501b8b982044f1c99d18e7dc62e6: Status 404 returned error can't find the container with id 805f4b013934c0f9ea85187e55df41599e25501b8b982044f1c99d18e7dc62e6 Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.126869 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.128714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.130574 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-82htt" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.131185 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.131320 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.131364 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.136298 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245104 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245149 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245185 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45lnx\" (UniqueName: \"kubernetes.io/projected/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kube-api-access-45lnx\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245273 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.245296 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.346944 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.346991 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.347033 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.347060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.347082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45lnx\" (UniqueName: \"kubernetes.io/projected/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kube-api-access-45lnx\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.347111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.347151 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.347183 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.348002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.348279 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.348531 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.349282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.349795 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.365067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.396184 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45lnx\" (UniqueName: \"kubernetes.io/projected/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kube-api-access-45lnx\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.396756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.418806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.422613 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.423646 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.429435 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jdfrt" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.430801 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.430918 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.443825 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.459254 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.562129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-combined-ca-bundle\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.562522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skkkn\" (UniqueName: \"kubernetes.io/projected/53c32262-ad33-4c11-9536-2b094ef77d55-kube-api-access-skkkn\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.563345 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-config-data\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.563592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-memcached-tls-certs\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.564066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-kolla-config\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.665350 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-combined-ca-bundle\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.665398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skkkn\" (UniqueName: \"kubernetes.io/projected/53c32262-ad33-4c11-9536-2b094ef77d55-kube-api-access-skkkn\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.665444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-config-data\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.665478 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-memcached-tls-certs\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.665548 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-kolla-config\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.666371 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-kolla-config\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.666603 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-config-data\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.669343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-memcached-tls-certs\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.669470 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-combined-ca-bundle\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.717175 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skkkn\" (UniqueName: \"kubernetes.io/projected/53c32262-ad33-4c11-9536-2b094ef77d55-kube-api-access-skkkn\") pod \"memcached-0\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " pod="openstack/memcached-0" Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.731533 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9","Type":"ContainerStarted","Data":"805f4b013934c0f9ea85187e55df41599e25501b8b982044f1c99d18e7dc62e6"} Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.766618 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"09f7d635e94f05ae8e8f470c73493c96f0ec6c0a4510de5411cdb6df882057b9"} Dec 05 11:28:15 crc kubenswrapper[4809]: I1205 11:28:15.771211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 11:28:16 crc kubenswrapper[4809]: I1205 11:28:16.097001 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:28:16 crc kubenswrapper[4809]: W1205 11:28:16.136999 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2b5c2a4_1802_4308_9db8_090ea111f7ba.slice/crio-f4e573e003cfef9d3f12853fee2d8c51a2d2a055035dfc75fef29ed7dcb19d04 WatchSource:0}: Error finding container f4e573e003cfef9d3f12853fee2d8c51a2d2a055035dfc75fef29ed7dcb19d04: Status 404 returned error can't find the container with id f4e573e003cfef9d3f12853fee2d8c51a2d2a055035dfc75fef29ed7dcb19d04 Dec 05 11:28:16 crc kubenswrapper[4809]: I1205 11:28:16.410072 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 11:28:16 crc kubenswrapper[4809]: W1205 11:28:16.468435 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53c32262_ad33_4c11_9536_2b094ef77d55.slice/crio-c86e0256b321bd4d1de49795129a9fa46476f82cf92d0e6ae1dca7b1142a75a7 WatchSource:0}: Error finding container c86e0256b321bd4d1de49795129a9fa46476f82cf92d0e6ae1dca7b1142a75a7: Status 404 returned error can't find the container with id c86e0256b321bd4d1de49795129a9fa46476f82cf92d0e6ae1dca7b1142a75a7 Dec 05 11:28:16 crc kubenswrapper[4809]: I1205 11:28:16.784557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d2b5c2a4-1802-4308-9db8-090ea111f7ba","Type":"ContainerStarted","Data":"f4e573e003cfef9d3f12853fee2d8c51a2d2a055035dfc75fef29ed7dcb19d04"} Dec 05 11:28:16 crc kubenswrapper[4809]: I1205 11:28:16.786269 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"53c32262-ad33-4c11-9536-2b094ef77d55","Type":"ContainerStarted","Data":"c86e0256b321bd4d1de49795129a9fa46476f82cf92d0e6ae1dca7b1142a75a7"} Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.581899 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.585264 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.590452 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4jnk8" Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.615205 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.728257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggrjx\" (UniqueName: \"kubernetes.io/projected/bcee5292-e219-4b51-871e-a283dc5cb231-kube-api-access-ggrjx\") pod \"kube-state-metrics-0\" (UID: \"bcee5292-e219-4b51-871e-a283dc5cb231\") " pod="openstack/kube-state-metrics-0" Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.829418 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggrjx\" (UniqueName: \"kubernetes.io/projected/bcee5292-e219-4b51-871e-a283dc5cb231-kube-api-access-ggrjx\") pod \"kube-state-metrics-0\" (UID: \"bcee5292-e219-4b51-871e-a283dc5cb231\") " pod="openstack/kube-state-metrics-0" Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.861737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggrjx\" (UniqueName: \"kubernetes.io/projected/bcee5292-e219-4b51-871e-a283dc5cb231-kube-api-access-ggrjx\") pod \"kube-state-metrics-0\" (UID: \"bcee5292-e219-4b51-871e-a283dc5cb231\") " pod="openstack/kube-state-metrics-0" Dec 05 11:28:17 crc kubenswrapper[4809]: I1205 11:28:17.932113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.866281 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fg4wz"] Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.868018 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.871212 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-s5vzd" Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.871519 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.871552 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.879575 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fg4wz"] Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.888996 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-bz8bw"] Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.893347 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:21 crc kubenswrapper[4809]: I1205 11:28:21.942557 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bz8bw"] Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015526 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11752f85-2262-477b-9920-bf5c13f45f20-scripts\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-run\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015652 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8j5c\" (UniqueName: \"kubernetes.io/projected/e8226c08-75fc-4250-bc4d-e0dbbc818730-kube-api-access-p8j5c\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-combined-ca-bundle\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015705 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8226c08-75fc-4250-bc4d-e0dbbc818730-scripts\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015760 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-log\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp5tq\" (UniqueName: \"kubernetes.io/projected/11752f85-2262-477b-9920-bf5c13f45f20-kube-api-access-mp5tq\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015844 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-etc-ovs\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015943 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-log-ovn\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.015996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-lib\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.016014 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run-ovn\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.016035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-ovn-controller-tls-certs\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.016066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.117800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11752f85-2262-477b-9920-bf5c13f45f20-scripts\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.117842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-run\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.117872 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8j5c\" (UniqueName: \"kubernetes.io/projected/e8226c08-75fc-4250-bc4d-e0dbbc818730-kube-api-access-p8j5c\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.117908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-combined-ca-bundle\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.117934 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8226c08-75fc-4250-bc4d-e0dbbc818730-scripts\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.117959 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-log\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.117979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp5tq\" (UniqueName: \"kubernetes.io/projected/11752f85-2262-477b-9920-bf5c13f45f20-kube-api-access-mp5tq\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-etc-ovs\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118049 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-log-ovn\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-lib\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118087 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run-ovn\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-ovn-controller-tls-certs\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118592 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-run\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-log-ovn\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118790 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-etc-ovs\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.118901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-lib\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.119030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run-ovn\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.119134 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-log\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.120109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11752f85-2262-477b-9920-bf5c13f45f20-scripts\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.120847 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8226c08-75fc-4250-bc4d-e0dbbc818730-scripts\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.126343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-combined-ca-bundle\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.139458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-ovn-controller-tls-certs\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.146584 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp5tq\" (UniqueName: \"kubernetes.io/projected/11752f85-2262-477b-9920-bf5c13f45f20-kube-api-access-mp5tq\") pod \"ovn-controller-ovs-bz8bw\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.147095 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8j5c\" (UniqueName: \"kubernetes.io/projected/e8226c08-75fc-4250-bc4d-e0dbbc818730-kube-api-access-p8j5c\") pod \"ovn-controller-fg4wz\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.201334 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:22 crc kubenswrapper[4809]: I1205 11:28:22.216935 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.678107 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.679611 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.682491 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.688518 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-bngtp" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.688668 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.689043 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.692598 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.694602 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.857255 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8434925e-b186-4558-956c-afa51f21113d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.858179 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.858234 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.858313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqnk8\" (UniqueName: \"kubernetes.io/projected/8434925e-b186-4558-956c-afa51f21113d-kube-api-access-cqnk8\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.858442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-config\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.858575 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.858671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.858737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.897758 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.904305 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.912292 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.915883 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.916105 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.916214 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.916584 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-m6bgq" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960561 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8434925e-b186-4558-956c-afa51f21113d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960602 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960659 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqnk8\" (UniqueName: \"kubernetes.io/projected/8434925e-b186-4558-956c-afa51f21113d-kube-api-access-cqnk8\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-config\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960795 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.960822 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.961135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8434925e-b186-4558-956c-afa51f21113d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.961514 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.962720 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.963000 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.964743 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.965041 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.969935 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.972129 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.972866 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-config\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.974582 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.975003 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.979274 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqnk8\" (UniqueName: \"kubernetes.io/projected/8434925e-b186-4558-956c-afa51f21113d-kube-api-access-cqnk8\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:24 crc kubenswrapper[4809]: I1205 11:28:24.981216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.002115 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-bngtp" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.010141 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-config\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbklf\" (UniqueName: \"kubernetes.io/projected/fd04fa46-c62b-4b8e-947d-da5846355feb-kube-api-access-fbklf\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062644 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062686 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.062812 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.164971 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-config\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.165040 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.165074 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.165140 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbklf\" (UniqueName: \"kubernetes.io/projected/fd04fa46-c62b-4b8e-947d-da5846355feb-kube-api-access-fbklf\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.165181 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.165213 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.165242 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.165278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.166053 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-config\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.166407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.166975 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.169219 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.172774 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.172903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.177296 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.187441 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbklf\" (UniqueName: \"kubernetes.io/projected/fd04fa46-c62b-4b8e-947d-da5846355feb-kube-api-access-fbklf\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.192004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:25 crc kubenswrapper[4809]: I1205 11:28:25.239811 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 11:28:41 crc kubenswrapper[4809]: E1205 11:28:41.100890 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 05 11:28:41 crc kubenswrapper[4809]: E1205 11:28:41.101426 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tncfv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(6b760a19-f9d4-4ea7-a6c7-6610aa0334a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:41 crc kubenswrapper[4809]: E1205 11:28:41.102650 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" Dec 05 11:28:41 crc kubenswrapper[4809]: E1205 11:28:41.968925 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.271880 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.272345 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sqxrg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(cfe522ab-449f-433e-a9bf-b33ed3b42595): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.273527 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.284762 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.285003 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-45lnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(d2b5c2a4-1802-4308-9db8-090ea111f7ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.286479 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.290846 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.291024 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rpxvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(e1251531-cb99-418e-aa38-65a793e5a9d0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.292628 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.994600 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.994877 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" Dec 05 11:28:45 crc kubenswrapper[4809]: E1205 11:28:45.994918 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.758132 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.758881 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dj9sw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-49vg4_openstack(f38368a6-2aa9-454d-bf49-816d959bc81c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.764311 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" podUID="f38368a6-2aa9-454d-bf49-816d959bc81c" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.784995 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.785328 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7lv4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-kd7lk_openstack(f5b74800-4cb0-4eb3-9d70-7b1935f49ca5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.790003 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" podUID="f5b74800-4cb0-4eb3-9d70-7b1935f49ca5" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.800589 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.800853 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kdjh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-zjbwc_openstack(c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.802120 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" podUID="c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.830450 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.832211 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jz4zx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-nm8sc_openstack(e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:28:51 crc kubenswrapper[4809]: E1205 11:28:51.835792 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" podUID="e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2" Dec 05 11:28:52 crc kubenswrapper[4809]: E1205 11:28:52.067130 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" podUID="f38368a6-2aa9-454d-bf49-816d959bc81c" Dec 05 11:28:52 crc kubenswrapper[4809]: E1205 11:28:52.067509 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" podUID="e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.253261 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fg4wz"] Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.394951 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bz8bw"] Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.558575 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.564514 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:52 crc kubenswrapper[4809]: W1205 11:28:52.567532 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcee5292_e219_4b51_871e_a283dc5cb231.slice/crio-c24383c5e8e7ea5ddf62d3c90c226f0f8352b4f73f54415b7ab5727885be4da2 WatchSource:0}: Error finding container c24383c5e8e7ea5ddf62d3c90c226f0f8352b4f73f54415b7ab5727885be4da2: Status 404 returned error can't find the container with id c24383c5e8e7ea5ddf62d3c90c226f0f8352b4f73f54415b7ab5727885be4da2 Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.591332 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.641927 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:28:52 crc kubenswrapper[4809]: W1205 11:28:52.646213 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8434925e_b186_4558_956c_afa51f21113d.slice/crio-b0dd5ca362f30df1e4de05c60facfdc7242d5f815d785284025b685fb7a6da4a WatchSource:0}: Error finding container b0dd5ca362f30df1e4de05c60facfdc7242d5f815d785284025b685fb7a6da4a: Status 404 returned error can't find the container with id b0dd5ca362f30df1e4de05c60facfdc7242d5f815d785284025b685fb7a6da4a Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.693833 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdjh8\" (UniqueName: \"kubernetes.io/projected/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-kube-api-access-kdjh8\") pod \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.693902 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-dns-svc\") pod \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.693936 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-config\") pod \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.694058 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lv4g\" (UniqueName: \"kubernetes.io/projected/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-kube-api-access-7lv4g\") pod \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\" (UID: \"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5\") " Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.694166 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-config\") pod \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\" (UID: \"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087\") " Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.694614 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087" (UID: "c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.695353 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-config" (OuterVolumeSpecName: "config") pod "f5b74800-4cb0-4eb3-9d70-7b1935f49ca5" (UID: "f5b74800-4cb0-4eb3-9d70-7b1935f49ca5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.695426 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-config" (OuterVolumeSpecName: "config") pod "c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087" (UID: "c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.699962 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-kube-api-access-kdjh8" (OuterVolumeSpecName: "kube-api-access-kdjh8") pod "c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087" (UID: "c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087"). InnerVolumeSpecName "kube-api-access-kdjh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.700369 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-kube-api-access-7lv4g" (OuterVolumeSpecName: "kube-api-access-7lv4g") pod "f5b74800-4cb0-4eb3-9d70-7b1935f49ca5" (UID: "f5b74800-4cb0-4eb3-9d70-7b1935f49ca5"). InnerVolumeSpecName "kube-api-access-7lv4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.796600 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.796665 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.796676 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lv4g\" (UniqueName: \"kubernetes.io/projected/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5-kube-api-access-7lv4g\") on node \"crc\" DevicePath \"\"" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.796688 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:28:52 crc kubenswrapper[4809]: I1205 11:28:52.796697 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdjh8\" (UniqueName: \"kubernetes.io/projected/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087-kube-api-access-kdjh8\") on node \"crc\" DevicePath \"\"" Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.068290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bz8bw" event={"ID":"11752f85-2262-477b-9920-bf5c13f45f20","Type":"ContainerStarted","Data":"9adf06c9ae9df7ca21d5ede2cfc3fc966b7d0e941944b2e08a6149a9954c264f"} Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.071666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz" event={"ID":"e8226c08-75fc-4250-bc4d-e0dbbc818730","Type":"ContainerStarted","Data":"8ef695dcb5b60d4deb4771eecf9462788c8a8b179c5543a1eb5405ec1f1e7ddf"} Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.075727 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8434925e-b186-4558-956c-afa51f21113d","Type":"ContainerStarted","Data":"b0dd5ca362f30df1e4de05c60facfdc7242d5f815d785284025b685fb7a6da4a"} Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.079301 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.079298 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zjbwc" event={"ID":"c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087","Type":"ContainerDied","Data":"9c02a4cb0427f5048ff7ccff44d9dce48e7ca3e24d8cf82cbaa759e182c665c6"} Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.080902 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bcee5292-e219-4b51-871e-a283dc5cb231","Type":"ContainerStarted","Data":"c24383c5e8e7ea5ddf62d3c90c226f0f8352b4f73f54415b7ab5727885be4da2"} Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.084134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"53c32262-ad33-4c11-9536-2b094ef77d55","Type":"ContainerStarted","Data":"df1855ca03cbe3f85a7e21b167634d60144ff55554d113b302ffc0afc2b270ce"} Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.084263 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.090799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" event={"ID":"f5b74800-4cb0-4eb3-9d70-7b1935f49ca5","Type":"ContainerDied","Data":"8b0cb67af22987e1fa76145856ab4978d99375841596bcf6df8f172c77267f03"} Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.090857 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd7lk" Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.108200 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.823337092 podStartE2EDuration="38.108178542s" podCreationTimestamp="2025-12-05 11:28:15 +0000 UTC" firstStartedPulling="2025-12-05 11:28:16.471763868 +0000 UTC m=+1191.862740426" lastFinishedPulling="2025-12-05 11:28:51.756605318 +0000 UTC m=+1227.147581876" observedRunningTime="2025-12-05 11:28:53.09739515 +0000 UTC m=+1228.488371708" watchObservedRunningTime="2025-12-05 11:28:53.108178542 +0000 UTC m=+1228.499155100" Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.137491 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd7lk"] Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.149527 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd7lk"] Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.160306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.176709 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zjbwc"] Dec 05 11:28:53 crc kubenswrapper[4809]: I1205 11:28:53.183540 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zjbwc"] Dec 05 11:28:54 crc kubenswrapper[4809]: I1205 11:28:54.099152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fd04fa46-c62b-4b8e-947d-da5846355feb","Type":"ContainerStarted","Data":"608ae4c80554477ca0ddaa85f9b68eadeb3d50c8750f03f07347c8727427fb4f"} Dec 05 11:28:54 crc kubenswrapper[4809]: I1205 11:28:54.881508 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087" path="/var/lib/kubelet/pods/c1a9f03c-dd4d-4fa7-9c9f-1c34724f8087/volumes" Dec 05 11:28:54 crc kubenswrapper[4809]: I1205 11:28:54.882533 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5b74800-4cb0-4eb3-9d70-7b1935f49ca5" path="/var/lib/kubelet/pods/f5b74800-4cb0-4eb3-9d70-7b1935f49ca5/volumes" Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.142083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8434925e-b186-4558-956c-afa51f21113d","Type":"ContainerStarted","Data":"0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a"} Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.144549 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bcee5292-e219-4b51-871e-a283dc5cb231","Type":"ContainerStarted","Data":"62fac1898d75b05fbf7bcbdeb2f3a3ddc669467f4255468d71872dab180ef4b4"} Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.145748 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.148024 4809 generic.go:334] "Generic (PLEG): container finished" podID="11752f85-2262-477b-9920-bf5c13f45f20" containerID="a62571ec4d894419275eeeda6a564866e0faca31c38feba3449fddfeb7025a82" exitCode=0 Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.148099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bz8bw" event={"ID":"11752f85-2262-477b-9920-bf5c13f45f20","Type":"ContainerDied","Data":"a62571ec4d894419275eeeda6a564866e0faca31c38feba3449fddfeb7025a82"} Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.149830 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz" event={"ID":"e8226c08-75fc-4250-bc4d-e0dbbc818730","Type":"ContainerStarted","Data":"6d5cb5ed862cbf7d2734d0573ce32c040186fb0fb53685e75922a348ec328b42"} Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.149901 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-fg4wz" Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.151594 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fd04fa46-c62b-4b8e-947d-da5846355feb","Type":"ContainerStarted","Data":"f754ccd26270d79dad0c1628da75aaa14922d40a673887d3fffac5fd152c8685"} Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.153688 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9","Type":"ContainerStarted","Data":"2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4"} Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.163557 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=36.487922177 podStartE2EDuration="42.163540954s" podCreationTimestamp="2025-12-05 11:28:17 +0000 UTC" firstStartedPulling="2025-12-05 11:28:52.572331591 +0000 UTC m=+1227.963308149" lastFinishedPulling="2025-12-05 11:28:58.247950368 +0000 UTC m=+1233.638926926" observedRunningTime="2025-12-05 11:28:59.15879579 +0000 UTC m=+1234.549772378" watchObservedRunningTime="2025-12-05 11:28:59.163540954 +0000 UTC m=+1234.554517512" Dec 05 11:28:59 crc kubenswrapper[4809]: I1205 11:28:59.227570 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fg4wz" podStartSLOduration=32.832588557 podStartE2EDuration="38.227547509s" podCreationTimestamp="2025-12-05 11:28:21 +0000 UTC" firstStartedPulling="2025-12-05 11:28:52.290448486 +0000 UTC m=+1227.681425054" lastFinishedPulling="2025-12-05 11:28:57.685407448 +0000 UTC m=+1233.076384006" observedRunningTime="2025-12-05 11:28:59.22719131 +0000 UTC m=+1234.618167878" watchObservedRunningTime="2025-12-05 11:28:59.227547509 +0000 UTC m=+1234.618524067" Dec 05 11:29:00 crc kubenswrapper[4809]: I1205 11:29:00.163991 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bz8bw" event={"ID":"11752f85-2262-477b-9920-bf5c13f45f20","Type":"ContainerStarted","Data":"5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4"} Dec 05 11:29:00 crc kubenswrapper[4809]: I1205 11:29:00.164476 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:29:00 crc kubenswrapper[4809]: I1205 11:29:00.164490 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:29:00 crc kubenswrapper[4809]: I1205 11:29:00.164498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bz8bw" event={"ID":"11752f85-2262-477b-9920-bf5c13f45f20","Type":"ContainerStarted","Data":"911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62"} Dec 05 11:29:00 crc kubenswrapper[4809]: I1205 11:29:00.165287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e1251531-cb99-418e-aa38-65a793e5a9d0","Type":"ContainerStarted","Data":"03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1"} Dec 05 11:29:00 crc kubenswrapper[4809]: I1205 11:29:00.183861 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-bz8bw" podStartSLOduration=34.068999728 podStartE2EDuration="39.183841491s" podCreationTimestamp="2025-12-05 11:28:21 +0000 UTC" firstStartedPulling="2025-12-05 11:28:52.408777442 +0000 UTC m=+1227.799754000" lastFinishedPulling="2025-12-05 11:28:57.523619195 +0000 UTC m=+1232.914595763" observedRunningTime="2025-12-05 11:29:00.180412282 +0000 UTC m=+1235.571388840" watchObservedRunningTime="2025-12-05 11:29:00.183841491 +0000 UTC m=+1235.574818049" Dec 05 11:29:00 crc kubenswrapper[4809]: I1205 11:29:00.773904 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 05 11:29:03 crc kubenswrapper[4809]: I1205 11:29:03.189694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fd04fa46-c62b-4b8e-947d-da5846355feb","Type":"ContainerStarted","Data":"08a2f85174278c931e6dfac745c0fe3ac459b86ac3c72c047078366ce8f1b112"} Dec 05 11:29:03 crc kubenswrapper[4809]: I1205 11:29:03.193276 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerID="2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4" exitCode=0 Dec 05 11:29:03 crc kubenswrapper[4809]: I1205 11:29:03.193344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9","Type":"ContainerDied","Data":"2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4"} Dec 05 11:29:03 crc kubenswrapper[4809]: I1205 11:29:03.196865 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8434925e-b186-4558-956c-afa51f21113d","Type":"ContainerStarted","Data":"17bf3535a2282730bf4584649cdef0fdc6c738363db31b9323fc6bba727833db"} Dec 05 11:29:03 crc kubenswrapper[4809]: I1205 11:29:03.199832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d2b5c2a4-1802-4308-9db8-090ea111f7ba","Type":"ContainerStarted","Data":"1f93774478558923bc3ec3b4f4d69eff36b9f1c081b7f2d5ab84f3d2a7b7c613"} Dec 05 11:29:03 crc kubenswrapper[4809]: I1205 11:29:03.230123 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=31.27267425 podStartE2EDuration="40.230099708s" podCreationTimestamp="2025-12-05 11:28:23 +0000 UTC" firstStartedPulling="2025-12-05 11:28:53.1455614 +0000 UTC m=+1228.536537948" lastFinishedPulling="2025-12-05 11:29:02.102986858 +0000 UTC m=+1237.493963406" observedRunningTime="2025-12-05 11:29:03.220551489 +0000 UTC m=+1238.611528047" watchObservedRunningTime="2025-12-05 11:29:03.230099708 +0000 UTC m=+1238.621076256" Dec 05 11:29:03 crc kubenswrapper[4809]: I1205 11:29:03.247185 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=30.80979 podStartE2EDuration="40.247156025s" podCreationTimestamp="2025-12-05 11:28:23 +0000 UTC" firstStartedPulling="2025-12-05 11:28:52.64870655 +0000 UTC m=+1228.039683108" lastFinishedPulling="2025-12-05 11:29:02.086072545 +0000 UTC m=+1237.477049133" observedRunningTime="2025-12-05 11:29:03.244511935 +0000 UTC m=+1238.635488493" watchObservedRunningTime="2025-12-05 11:29:03.247156025 +0000 UTC m=+1238.638132583" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.011705 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.051450 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.211206 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9","Type":"ContainerStarted","Data":"cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f"} Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.212939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfe522ab-449f-433e-a9bf-b33ed3b42595","Type":"ContainerStarted","Data":"406129ebd40a9895b56eb4df3cb21847834b08c83a4aab01708e82905b319f2a"} Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.213538 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.236872 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.006089721 podStartE2EDuration="52.23683756s" podCreationTimestamp="2025-12-05 11:28:12 +0000 UTC" firstStartedPulling="2025-12-05 11:28:14.933183681 +0000 UTC m=+1190.324160239" lastFinishedPulling="2025-12-05 11:28:58.16393152 +0000 UTC m=+1233.554908078" observedRunningTime="2025-12-05 11:29:04.231800049 +0000 UTC m=+1239.622776627" watchObservedRunningTime="2025-12-05 11:29:04.23683756 +0000 UTC m=+1239.627814108" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.240124 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.257118 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.259102 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.259126 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.280626 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.469527 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nm8sc"] Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.525886 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-94zqc"] Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.529149 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.533288 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.545914 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gpkvt"] Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.549492 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.556222 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.557913 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-94zqc"] Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.573810 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gpkvt"] Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.631542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a2c6a1-9783-4742-948a-d992416fd6cc-config\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.632137 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovn-rundir\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.632422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.632689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovs-rundir\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.632718 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-combined-ca-bundle\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.632758 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zv92\" (UniqueName: \"kubernetes.io/projected/a5a2c6a1-9783-4742-948a-d992416fd6cc-kube-api-access-9zv92\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.737417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.737559 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk5jz\" (UniqueName: \"kubernetes.io/projected/ae470489-d955-443f-a16b-aa09f3955cd2-kube-api-access-sk5jz\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.737592 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovs-rundir\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.737660 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-combined-ca-bundle\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.737727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zv92\" (UniqueName: \"kubernetes.io/projected/a5a2c6a1-9783-4742-948a-d992416fd6cc-kube-api-access-9zv92\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.737759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-config\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.738013 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.738226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a2c6a1-9783-4742-948a-d992416fd6cc-config\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.738252 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovn-rundir\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.738303 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.739439 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovs-rundir\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.740456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovn-rundir\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.740551 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a2c6a1-9783-4742-948a-d992416fd6cc-config\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.745144 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-combined-ca-bundle\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.754378 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.766755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zv92\" (UniqueName: \"kubernetes.io/projected/a5a2c6a1-9783-4742-948a-d992416fd6cc-kube-api-access-9zv92\") pod \"ovn-controller-metrics-94zqc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.841110 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.841797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk5jz\" (UniqueName: \"kubernetes.io/projected/ae470489-d955-443f-a16b-aa09f3955cd2-kube-api-access-sk5jz\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.841896 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-config\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.841954 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.843000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.843897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-config\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.844808 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.858172 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.864917 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.868458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk5jz\" (UniqueName: \"kubernetes.io/projected/ae470489-d955-443f-a16b-aa09f3955cd2-kube-api-access-sk5jz\") pod \"dnsmasq-dns-7fd796d7df-gpkvt\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.877350 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.936067 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-49vg4"] Dec 05 11:29:04 crc kubenswrapper[4809]: I1205 11:29:04.966614 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8shnb"] Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.008120 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.014343 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.022934 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8shnb"] Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.044476 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-dns-svc\") pod \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.044543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz4zx\" (UniqueName: \"kubernetes.io/projected/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-kube-api-access-jz4zx\") pod \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.044692 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-config\") pod \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\" (UID: \"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2\") " Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.045175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2" (UID: "e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.045542 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-config" (OuterVolumeSpecName: "config") pod "e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2" (UID: "e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.063985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-kube-api-access-jz4zx" (OuterVolumeSpecName: "kube-api-access-jz4zx") pod "e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2" (UID: "e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2"). InnerVolumeSpecName "kube-api-access-jz4zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.148734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.149184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.149220 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vddz\" (UniqueName: \"kubernetes.io/projected/7d49f737-1a41-4f92-b47e-d2935151a71c-kube-api-access-2vddz\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.149282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-config\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.149307 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.149390 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.149413 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.149426 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz4zx\" (UniqueName: \"kubernetes.io/projected/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2-kube-api-access-jz4zx\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.222104 4809 generic.go:334] "Generic (PLEG): container finished" podID="f38368a6-2aa9-454d-bf49-816d959bc81c" containerID="cd2b0f11096bc64b41f94a16ef5f5a73adf7215119a099af8ffe3f311c7d62b2" exitCode=0 Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.222161 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" event={"ID":"f38368a6-2aa9-454d-bf49-816d959bc81c","Type":"ContainerDied","Data":"cd2b0f11096bc64b41f94a16ef5f5a73adf7215119a099af8ffe3f311c7d62b2"} Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.224592 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.229397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nm8sc" event={"ID":"e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2","Type":"ContainerDied","Data":"5ec9778ceb64b25ebb773904ab9a314cce960c967b771d01e6ea99224f300080"} Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.229774 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.252849 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-config\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.252894 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.252970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.253032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.253049 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vddz\" (UniqueName: \"kubernetes.io/projected/7d49f737-1a41-4f92-b47e-d2935151a71c-kube-api-access-2vddz\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.253879 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-config\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.254263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.254791 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.255067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.275290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vddz\" (UniqueName: \"kubernetes.io/projected/7d49f737-1a41-4f92-b47e-d2935151a71c-kube-api-access-2vddz\") pod \"dnsmasq-dns-86db49b7ff-8shnb\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.303041 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.304371 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nm8sc"] Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.312118 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nm8sc"] Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.394537 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.437466 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-94zqc"] Dec 05 11:29:05 crc kubenswrapper[4809]: W1205 11:29:05.574821 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae470489_d955_443f_a16b_aa09f3955cd2.slice/crio-5a91a0a90519b7897a8eac84d91c5dbc10ba5f23819d23119e4b529fa7abecf6 WatchSource:0}: Error finding container 5a91a0a90519b7897a8eac84d91c5dbc10ba5f23819d23119e4b529fa7abecf6: Status 404 returned error can't find the container with id 5a91a0a90519b7897a8eac84d91c5dbc10ba5f23819d23119e4b529fa7abecf6 Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.586892 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gpkvt"] Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.656763 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.662110 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.668305 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.668460 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.668305 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.669303 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-vvbwd" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.671363 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.693072 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778393 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj9sw\" (UniqueName: \"kubernetes.io/projected/f38368a6-2aa9-454d-bf49-816d959bc81c-kube-api-access-dj9sw\") pod \"f38368a6-2aa9-454d-bf49-816d959bc81c\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778434 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-dns-svc\") pod \"f38368a6-2aa9-454d-bf49-816d959bc81c\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778456 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-config\") pod \"f38368a6-2aa9-454d-bf49-816d959bc81c\" (UID: \"f38368a6-2aa9-454d-bf49-816d959bc81c\") " Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778750 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-config\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778808 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbw6c\" (UniqueName: \"kubernetes.io/projected/3aabcdcf-4f51-4398-ae40-33b6d57aa514-kube-api-access-sbw6c\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778842 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-scripts\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.778897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.786531 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f38368a6-2aa9-454d-bf49-816d959bc81c-kube-api-access-dj9sw" (OuterVolumeSpecName: "kube-api-access-dj9sw") pod "f38368a6-2aa9-454d-bf49-816d959bc81c" (UID: "f38368a6-2aa9-454d-bf49-816d959bc81c"). InnerVolumeSpecName "kube-api-access-dj9sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.817317 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-config" (OuterVolumeSpecName: "config") pod "f38368a6-2aa9-454d-bf49-816d959bc81c" (UID: "f38368a6-2aa9-454d-bf49-816d959bc81c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.826405 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f38368a6-2aa9-454d-bf49-816d959bc81c" (UID: "f38368a6-2aa9-454d-bf49-816d959bc81c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.880452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-scripts\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.880960 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.880993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-config\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881113 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbw6c\" (UniqueName: \"kubernetes.io/projected/3aabcdcf-4f51-4398-ae40-33b6d57aa514-kube-api-access-sbw6c\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881167 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj9sw\" (UniqueName: \"kubernetes.io/projected/f38368a6-2aa9-454d-bf49-816d959bc81c-kube-api-access-dj9sw\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881181 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881193 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38368a6-2aa9-454d-bf49-816d959bc81c-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.881810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-scripts\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.882218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.886591 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-config\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.895657 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.895706 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.895985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.929712 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbw6c\" (UniqueName: \"kubernetes.io/projected/3aabcdcf-4f51-4398-ae40-33b6d57aa514-kube-api-access-sbw6c\") pod \"ovn-northd-0\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " pod="openstack/ovn-northd-0" Dec 05 11:29:05 crc kubenswrapper[4809]: I1205 11:29:05.989262 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8shnb"] Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.004567 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.234219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" event={"ID":"7d49f737-1a41-4f92-b47e-d2935151a71c","Type":"ContainerStarted","Data":"b189039130105de78915b07ae0163f1ed73b92a55d08d33e2692e02c7760b398"} Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.234476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" event={"ID":"7d49f737-1a41-4f92-b47e-d2935151a71c","Type":"ContainerStarted","Data":"8add03e920e8ac974244bf14db514f9bb082e851dda7c82c8d22ef71334aeeb6"} Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.236393 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" event={"ID":"f38368a6-2aa9-454d-bf49-816d959bc81c","Type":"ContainerDied","Data":"d14b51c5a0cf76b5c728ee1c78e1d4b66ea0b6f7ef968d5d3e3aaa44bcd31811"} Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.236422 4809 scope.go:117] "RemoveContainer" containerID="cd2b0f11096bc64b41f94a16ef5f5a73adf7215119a099af8ffe3f311c7d62b2" Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.236439 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-49vg4" Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.240188 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae470489-d955-443f-a16b-aa09f3955cd2" containerID="81740b7bb06abef1fdcbd7f012b5175aeb2e5a55ec5fe9307d510e2cb9987c43" exitCode=0 Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.240262 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" event={"ID":"ae470489-d955-443f-a16b-aa09f3955cd2","Type":"ContainerDied","Data":"81740b7bb06abef1fdcbd7f012b5175aeb2e5a55ec5fe9307d510e2cb9987c43"} Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.240286 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" event={"ID":"ae470489-d955-443f-a16b-aa09f3955cd2","Type":"ContainerStarted","Data":"5a91a0a90519b7897a8eac84d91c5dbc10ba5f23819d23119e4b529fa7abecf6"} Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.243031 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-94zqc" event={"ID":"a5a2c6a1-9783-4742-948a-d992416fd6cc","Type":"ContainerStarted","Data":"f41b7ad67c7562f28f0be7b6223e95bfb20a2e91b0355507be02a8e171f2fb04"} Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.243155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-94zqc" event={"ID":"a5a2c6a1-9783-4742-948a-d992416fd6cc","Type":"ContainerStarted","Data":"bc480c77184046a20419642141ef662d97a30ddf1835770d77412fdb147cf956"} Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.276651 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-94zqc" podStartSLOduration=2.276615142 podStartE2EDuration="2.276615142s" podCreationTimestamp="2025-12-05 11:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:06.269757883 +0000 UTC m=+1241.660734441" watchObservedRunningTime="2025-12-05 11:29:06.276615142 +0000 UTC m=+1241.667591700" Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.361404 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-49vg4"] Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.368660 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-49vg4"] Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.441155 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:29:06 crc kubenswrapper[4809]: W1205 11:29:06.449898 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aabcdcf_4f51_4398_ae40_33b6d57aa514.slice/crio-83d8e4cc27af30f9bc9fc36275cec53b7f0ff60bd01e0174cda8f42480d2b5bc WatchSource:0}: Error finding container 83d8e4cc27af30f9bc9fc36275cec53b7f0ff60bd01e0174cda8f42480d2b5bc: Status 404 returned error can't find the container with id 83d8e4cc27af30f9bc9fc36275cec53b7f0ff60bd01e0174cda8f42480d2b5bc Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.885947 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2" path="/var/lib/kubelet/pods/e5a72a5a-f9ed-483c-bc9d-1866d9a29ce2/volumes" Dec 05 11:29:06 crc kubenswrapper[4809]: I1205 11:29:06.886759 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f38368a6-2aa9-454d-bf49-816d959bc81c" path="/var/lib/kubelet/pods/f38368a6-2aa9-454d-bf49-816d959bc81c/volumes" Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.251824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" event={"ID":"ae470489-d955-443f-a16b-aa09f3955cd2","Type":"ContainerStarted","Data":"c03271fabaf817edeee1037ab92f5ab0ce5d687203c8c444c8e65693bcb7af60"} Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.251913 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.253209 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerID="b189039130105de78915b07ae0163f1ed73b92a55d08d33e2692e02c7760b398" exitCode=0 Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.253235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" event={"ID":"7d49f737-1a41-4f92-b47e-d2935151a71c","Type":"ContainerDied","Data":"b189039130105de78915b07ae0163f1ed73b92a55d08d33e2692e02c7760b398"} Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.257256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3aabcdcf-4f51-4398-ae40-33b6d57aa514","Type":"ContainerStarted","Data":"83d8e4cc27af30f9bc9fc36275cec53b7f0ff60bd01e0174cda8f42480d2b5bc"} Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.260048 4809 generic.go:334] "Generic (PLEG): container finished" podID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerID="1f93774478558923bc3ec3b4f4d69eff36b9f1c081b7f2d5ab84f3d2a7b7c613" exitCode=0 Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.260934 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d2b5c2a4-1802-4308-9db8-090ea111f7ba","Type":"ContainerDied","Data":"1f93774478558923bc3ec3b4f4d69eff36b9f1c081b7f2d5ab84f3d2a7b7c613"} Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.294333 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" podStartSLOduration=3.29429455 podStartE2EDuration="3.29429455s" podCreationTimestamp="2025-12-05 11:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:07.273398204 +0000 UTC m=+1242.664374772" watchObservedRunningTime="2025-12-05 11:29:07.29429455 +0000 UTC m=+1242.685271108" Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.881687 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gpkvt"] Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.917917 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-68v9p"] Dec 05 11:29:07 crc kubenswrapper[4809]: E1205 11:29:07.918236 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38368a6-2aa9-454d-bf49-816d959bc81c" containerName="init" Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.918249 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38368a6-2aa9-454d-bf49-816d959bc81c" containerName="init" Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.918404 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38368a6-2aa9-454d-bf49-816d959bc81c" containerName="init" Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.919189 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.934611 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-68v9p"] Dec 05 11:29:07 crc kubenswrapper[4809]: I1205 11:29:07.949300 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.031692 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.031752 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-config\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.031804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.031945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsxdg\" (UniqueName: \"kubernetes.io/projected/de639392-c7d7-4b7f-b894-0fb5d32f00a2-kube-api-access-xsxdg\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.031989 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.133569 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsxdg\" (UniqueName: \"kubernetes.io/projected/de639392-c7d7-4b7f-b894-0fb5d32f00a2-kube-api-access-xsxdg\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.133646 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.133695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.133711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-config\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.133743 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.134688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.134747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.134908 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.134930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-config\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.152938 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsxdg\" (UniqueName: \"kubernetes.io/projected/de639392-c7d7-4b7f-b894-0fb5d32f00a2-kube-api-access-xsxdg\") pod \"dnsmasq-dns-698758b865-68v9p\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.234171 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.667490 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-68v9p"] Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.957726 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.964572 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.969912 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.970228 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.970324 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-f7xbc" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.971677 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 05 11:29:08 crc kubenswrapper[4809]: I1205 11:29:08.976511 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.048294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.048732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.048803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-lock\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.048837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrb7j\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-kube-api-access-mrb7j\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.048917 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-cache\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.150933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-cache\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.151000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.151062 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.151134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-lock\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.151170 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrb7j\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-kube-api-access-mrb7j\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.152112 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-cache\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: E1205 11:29:09.152220 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:29:09 crc kubenswrapper[4809]: E1205 11:29:09.152241 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:29:09 crc kubenswrapper[4809]: E1205 11:29:09.152287 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift podName:58501ab1-18d7-45d9-98cb-a74e62db36fc nodeName:}" failed. No retries permitted until 2025-12-05 11:29:09.652267756 +0000 UTC m=+1245.043244324 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift") pod "swift-storage-0" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc") : configmap "swift-ring-files" not found Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.152558 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.153403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-lock\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.175341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrb7j\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-kube-api-access-mrb7j\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.181292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.280115 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-68v9p" event={"ID":"de639392-c7d7-4b7f-b894-0fb5d32f00a2","Type":"ContainerStarted","Data":"e8595373053c2406cac8fbb2231f57767a9303d2feda538df1853c6a7c06cfe8"} Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.280591 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" containerName="dnsmasq-dns" containerID="cri-o://c03271fabaf817edeee1037ab92f5ab0ce5d687203c8c444c8e65693bcb7af60" gracePeriod=10 Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.503137 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-tcsq6"] Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.504515 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.510586 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.515046 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.515191 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.525804 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tcsq6"] Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.560588 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-combined-ca-bundle\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.560652 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac85b554-c4a8-483f-a64d-cedf0984140e-etc-swift\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.560853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-scripts\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.560919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-swiftconf\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.560975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ml6t\" (UniqueName: \"kubernetes.io/projected/ac85b554-c4a8-483f-a64d-cedf0984140e-kube-api-access-5ml6t\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.561109 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-ring-data-devices\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.561172 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-dispersionconf\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.662706 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-ring-data-devices\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.662773 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-dispersionconf\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.662925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.662960 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-combined-ca-bundle\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.662993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac85b554-c4a8-483f-a64d-cedf0984140e-etc-swift\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.663052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-scripts\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.663090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-swiftconf\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: E1205 11:29:09.663104 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:29:09 crc kubenswrapper[4809]: E1205 11:29:09.663123 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:29:09 crc kubenswrapper[4809]: E1205 11:29:09.663170 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift podName:58501ab1-18d7-45d9-98cb-a74e62db36fc nodeName:}" failed. No retries permitted until 2025-12-05 11:29:10.663154864 +0000 UTC m=+1246.054131422 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift") pod "swift-storage-0" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc") : configmap "swift-ring-files" not found Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.663798 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac85b554-c4a8-483f-a64d-cedf0984140e-etc-swift\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.664076 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-ring-data-devices\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.664115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-scripts\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.664197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ml6t\" (UniqueName: \"kubernetes.io/projected/ac85b554-c4a8-483f-a64d-cedf0984140e-kube-api-access-5ml6t\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.667151 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-swiftconf\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.667265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-dispersionconf\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.667786 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-combined-ca-bundle\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.683507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ml6t\" (UniqueName: \"kubernetes.io/projected/ac85b554-c4a8-483f-a64d-cedf0984140e-kube-api-access-5ml6t\") pod \"swift-ring-rebalance-tcsq6\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:09 crc kubenswrapper[4809]: I1205 11:29:09.831580 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:10 crc kubenswrapper[4809]: I1205 11:29:10.311088 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tcsq6"] Dec 05 11:29:10 crc kubenswrapper[4809]: W1205 11:29:10.316591 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac85b554_c4a8_483f_a64d_cedf0984140e.slice/crio-0acd9c9e86a3931b37f4284e1c2feffafdcdc7206669e36d03e2b7b44a9c25dd WatchSource:0}: Error finding container 0acd9c9e86a3931b37f4284e1c2feffafdcdc7206669e36d03e2b7b44a9c25dd: Status 404 returned error can't find the container with id 0acd9c9e86a3931b37f4284e1c2feffafdcdc7206669e36d03e2b7b44a9c25dd Dec 05 11:29:10 crc kubenswrapper[4809]: I1205 11:29:10.685329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:10 crc kubenswrapper[4809]: E1205 11:29:10.685589 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:29:10 crc kubenswrapper[4809]: E1205 11:29:10.686001 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:29:10 crc kubenswrapper[4809]: E1205 11:29:10.686101 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift podName:58501ab1-18d7-45d9-98cb-a74e62db36fc nodeName:}" failed. No retries permitted until 2025-12-05 11:29:12.686079139 +0000 UTC m=+1248.077055697 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift") pod "swift-storage-0" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc") : configmap "swift-ring-files" not found Dec 05 11:29:11 crc kubenswrapper[4809]: I1205 11:29:11.298835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcsq6" event={"ID":"ac85b554-c4a8-483f-a64d-cedf0984140e","Type":"ContainerStarted","Data":"0acd9c9e86a3931b37f4284e1c2feffafdcdc7206669e36d03e2b7b44a9c25dd"} Dec 05 11:29:12 crc kubenswrapper[4809]: I1205 11:29:12.721371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:12 crc kubenswrapper[4809]: E1205 11:29:12.721547 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:29:12 crc kubenswrapper[4809]: E1205 11:29:12.722950 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:29:12 crc kubenswrapper[4809]: E1205 11:29:12.723024 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift podName:58501ab1-18d7-45d9-98cb-a74e62db36fc nodeName:}" failed. No retries permitted until 2025-12-05 11:29:16.723006386 +0000 UTC m=+1252.113982944 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift") pod "swift-storage-0" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc") : configmap "swift-ring-files" not found Dec 05 11:29:12 crc kubenswrapper[4809]: I1205 11:29:12.957550 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.176035 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.334367 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" event={"ID":"7d49f737-1a41-4f92-b47e-d2935151a71c","Type":"ContainerStarted","Data":"191de9b1fc5f9c1b4fce32c5d2c9973de7db04c83f0f2a9a384bbd4821136087"} Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.334650 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.346788 4809 generic.go:334] "Generic (PLEG): container finished" podID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerID="4501f7a7e855fa971938992a20e039bcf63b75bfe786a79b79d5ebde13b721ba" exitCode=0 Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.346883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-68v9p" event={"ID":"de639392-c7d7-4b7f-b894-0fb5d32f00a2","Type":"ContainerDied","Data":"4501f7a7e855fa971938992a20e039bcf63b75bfe786a79b79d5ebde13b721ba"} Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.350001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d2b5c2a4-1802-4308-9db8-090ea111f7ba","Type":"ContainerStarted","Data":"54582cab87c46307e3ea258499e465a65a3e2466f31df5e0ba4e63028feffc0f"} Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.352380 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae470489-d955-443f-a16b-aa09f3955cd2" containerID="c03271fabaf817edeee1037ab92f5ab0ce5d687203c8c444c8e65693bcb7af60" exitCode=0 Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.353219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" event={"ID":"ae470489-d955-443f-a16b-aa09f3955cd2","Type":"ContainerDied","Data":"c03271fabaf817edeee1037ab92f5ab0ce5d687203c8c444c8e65693bcb7af60"} Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.375673 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" podStartSLOduration=9.375655013 podStartE2EDuration="9.375655013s" podCreationTimestamp="2025-12-05 11:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:13.358414672 +0000 UTC m=+1248.749391250" watchObservedRunningTime="2025-12-05 11:29:13.375655013 +0000 UTC m=+1248.766631591" Dec 05 11:29:13 crc kubenswrapper[4809]: I1205 11:29:13.391116 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371977.463684 podStartE2EDuration="59.391091297s" podCreationTimestamp="2025-12-05 11:28:14 +0000 UTC" firstStartedPulling="2025-12-05 11:28:16.151038656 +0000 UTC m=+1191.542015224" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:13.385928531 +0000 UTC m=+1248.776905089" watchObservedRunningTime="2025-12-05 11:29:13.391091297 +0000 UTC m=+1248.782067855" Dec 05 11:29:14 crc kubenswrapper[4809]: I1205 11:29:14.363470 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3aabcdcf-4f51-4398-ae40-33b6d57aa514","Type":"ContainerStarted","Data":"043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328"} Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.373739 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-838e-account-create-update-rt9bd"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.374914 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.377158 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.386869 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-838e-account-create-update-rt9bd"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.439553 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lxhkk"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.440728 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.449419 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lxhkk"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.459413 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.459756 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.482863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcf740a-2a63-438c-8cd9-774324492a8e-operator-scripts\") pod \"keystone-838e-account-create-update-rt9bd\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.482972 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-operator-scripts\") pod \"keystone-db-create-lxhkk\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.483125 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m97td\" (UniqueName: \"kubernetes.io/projected/afcf740a-2a63-438c-8cd9-774324492a8e-kube-api-access-m97td\") pod \"keystone-838e-account-create-update-rt9bd\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.483170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r58qw\" (UniqueName: \"kubernetes.io/projected/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-kube-api-access-r58qw\") pod \"keystone-db-create-lxhkk\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.584247 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcf740a-2a63-438c-8cd9-774324492a8e-operator-scripts\") pod \"keystone-838e-account-create-update-rt9bd\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.584327 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-operator-scripts\") pod \"keystone-db-create-lxhkk\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.584469 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m97td\" (UniqueName: \"kubernetes.io/projected/afcf740a-2a63-438c-8cd9-774324492a8e-kube-api-access-m97td\") pod \"keystone-838e-account-create-update-rt9bd\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.584500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r58qw\" (UniqueName: \"kubernetes.io/projected/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-kube-api-access-r58qw\") pod \"keystone-db-create-lxhkk\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.585210 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcf740a-2a63-438c-8cd9-774324492a8e-operator-scripts\") pod \"keystone-838e-account-create-update-rt9bd\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.585292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-operator-scripts\") pod \"keystone-db-create-lxhkk\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.606290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m97td\" (UniqueName: \"kubernetes.io/projected/afcf740a-2a63-438c-8cd9-774324492a8e-kube-api-access-m97td\") pod \"keystone-838e-account-create-update-rt9bd\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.607017 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r58qw\" (UniqueName: \"kubernetes.io/projected/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-kube-api-access-r58qw\") pod \"keystone-db-create-lxhkk\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.635703 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-jllxv"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.637026 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.665352 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jllxv"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.686298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdtjd\" (UniqueName: \"kubernetes.io/projected/3aede7af-428a-4d18-b78d-39e2c438d140-kube-api-access-pdtjd\") pod \"placement-db-create-jllxv\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.686477 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aede7af-428a-4d18-b78d-39e2c438d140-operator-scripts\") pod \"placement-db-create-jllxv\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.695281 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.738504 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ef3a-account-create-update-9mdwj"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.739809 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.746226 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.747220 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ef3a-account-create-update-9mdwj"] Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.764007 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.788655 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22xnz\" (UniqueName: \"kubernetes.io/projected/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-kube-api-access-22xnz\") pod \"placement-ef3a-account-create-update-9mdwj\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.788754 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdtjd\" (UniqueName: \"kubernetes.io/projected/3aede7af-428a-4d18-b78d-39e2c438d140-kube-api-access-pdtjd\") pod \"placement-db-create-jllxv\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.788822 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-operator-scripts\") pod \"placement-ef3a-account-create-update-9mdwj\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.788914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aede7af-428a-4d18-b78d-39e2c438d140-operator-scripts\") pod \"placement-db-create-jllxv\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.789528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aede7af-428a-4d18-b78d-39e2c438d140-operator-scripts\") pod \"placement-db-create-jllxv\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.812543 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdtjd\" (UniqueName: \"kubernetes.io/projected/3aede7af-428a-4d18-b78d-39e2c438d140-kube-api-access-pdtjd\") pod \"placement-db-create-jllxv\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.889856 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22xnz\" (UniqueName: \"kubernetes.io/projected/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-kube-api-access-22xnz\") pod \"placement-ef3a-account-create-update-9mdwj\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.889955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-operator-scripts\") pod \"placement-ef3a-account-create-update-9mdwj\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.890805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-operator-scripts\") pod \"placement-ef3a-account-create-update-9mdwj\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.910689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22xnz\" (UniqueName: \"kubernetes.io/projected/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-kube-api-access-22xnz\") pod \"placement-ef3a-account-create-update-9mdwj\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.988179 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jllxv" Dec 05 11:29:15 crc kubenswrapper[4809]: I1205 11:29:15.998094 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.065233 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.092573 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-dns-svc\") pod \"ae470489-d955-443f-a16b-aa09f3955cd2\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.092716 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk5jz\" (UniqueName: \"kubernetes.io/projected/ae470489-d955-443f-a16b-aa09f3955cd2-kube-api-access-sk5jz\") pod \"ae470489-d955-443f-a16b-aa09f3955cd2\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.092788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-ovsdbserver-nb\") pod \"ae470489-d955-443f-a16b-aa09f3955cd2\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.092815 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-config\") pod \"ae470489-d955-443f-a16b-aa09f3955cd2\" (UID: \"ae470489-d955-443f-a16b-aa09f3955cd2\") " Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.098765 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae470489-d955-443f-a16b-aa09f3955cd2-kube-api-access-sk5jz" (OuterVolumeSpecName: "kube-api-access-sk5jz") pod "ae470489-d955-443f-a16b-aa09f3955cd2" (UID: "ae470489-d955-443f-a16b-aa09f3955cd2"). InnerVolumeSpecName "kube-api-access-sk5jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.130821 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-config" (OuterVolumeSpecName: "config") pod "ae470489-d955-443f-a16b-aa09f3955cd2" (UID: "ae470489-d955-443f-a16b-aa09f3955cd2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.132525 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae470489-d955-443f-a16b-aa09f3955cd2" (UID: "ae470489-d955-443f-a16b-aa09f3955cd2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.136940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae470489-d955-443f-a16b-aa09f3955cd2" (UID: "ae470489-d955-443f-a16b-aa09f3955cd2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.195137 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk5jz\" (UniqueName: \"kubernetes.io/projected/ae470489-d955-443f-a16b-aa09f3955cd2-kube-api-access-sk5jz\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.195173 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.195182 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.195191 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae470489-d955-443f-a16b-aa09f3955cd2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.384586 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.384577 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" event={"ID":"ae470489-d955-443f-a16b-aa09f3955cd2","Type":"ContainerDied","Data":"5a91a0a90519b7897a8eac84d91c5dbc10ba5f23819d23119e4b529fa7abecf6"} Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.384728 4809 scope.go:117] "RemoveContainer" containerID="c03271fabaf817edeee1037ab92f5ab0ce5d687203c8c444c8e65693bcb7af60" Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.415297 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gpkvt"] Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.420684 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gpkvt"] Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.807050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:16 crc kubenswrapper[4809]: E1205 11:29:16.807289 4809 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:29:16 crc kubenswrapper[4809]: E1205 11:29:16.807314 4809 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:29:16 crc kubenswrapper[4809]: E1205 11:29:16.807378 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift podName:58501ab1-18d7-45d9-98cb-a74e62db36fc nodeName:}" failed. No retries permitted until 2025-12-05 11:29:24.807356606 +0000 UTC m=+1260.198333164 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift") pod "swift-storage-0" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc") : configmap "swift-ring-files" not found Dec 05 11:29:16 crc kubenswrapper[4809]: I1205 11:29:16.881853 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" path="/var/lib/kubelet/pods/ae470489-d955-443f-a16b-aa09f3955cd2/volumes" Dec 05 11:29:17 crc kubenswrapper[4809]: I1205 11:29:17.070201 4809 scope.go:117] "RemoveContainer" containerID="81740b7bb06abef1fdcbd7f012b5175aeb2e5a55ec5fe9307d510e2cb9987c43" Dec 05 11:29:17 crc kubenswrapper[4809]: I1205 11:29:17.688677 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lxhkk"] Dec 05 11:29:17 crc kubenswrapper[4809]: I1205 11:29:17.711696 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-838e-account-create-update-rt9bd"] Dec 05 11:29:17 crc kubenswrapper[4809]: I1205 11:29:17.729530 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ef3a-account-create-update-9mdwj"] Dec 05 11:29:17 crc kubenswrapper[4809]: I1205 11:29:17.801270 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jllxv"] Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.414264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-68v9p" event={"ID":"de639392-c7d7-4b7f-b894-0fb5d32f00a2","Type":"ContainerStarted","Data":"e1ef014dab365de170de8bcc18bc67b0c2a056e8f731e71b7a9f0c26d47e7122"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.414570 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.416325 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcsq6" event={"ID":"ac85b554-c4a8-483f-a64d-cedf0984140e","Type":"ContainerStarted","Data":"bebae10222675f498e6edbdca8558f09457b6b61eba7576c606fad86d41f948e"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.419051 4809 generic.go:334] "Generic (PLEG): container finished" podID="afcf740a-2a63-438c-8cd9-774324492a8e" containerID="7387e0246accba106019ec8b535f729502b108f72a360c3a7ec7e0a7c8203dbd" exitCode=0 Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.419108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-838e-account-create-update-rt9bd" event={"ID":"afcf740a-2a63-438c-8cd9-774324492a8e","Type":"ContainerDied","Data":"7387e0246accba106019ec8b535f729502b108f72a360c3a7ec7e0a7c8203dbd"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.419128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-838e-account-create-update-rt9bd" event={"ID":"afcf740a-2a63-438c-8cd9-774324492a8e","Type":"ContainerStarted","Data":"d7dc30602b6526c2c13a5a47b8192b5defa06b965d3fdc1028fca43e5c70cf88"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.421384 4809 generic.go:334] "Generic (PLEG): container finished" podID="d074606c-4b4c-46bc-8552-d9a1ddcd9cc5" containerID="ca17fb3b9f7c4840c3062e70874e1e17ae9c14450b3a2c90ff5e4142c3b55b40" exitCode=0 Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.421548 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ef3a-account-create-update-9mdwj" event={"ID":"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5","Type":"ContainerDied","Data":"ca17fb3b9f7c4840c3062e70874e1e17ae9c14450b3a2c90ff5e4142c3b55b40"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.421595 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ef3a-account-create-update-9mdwj" event={"ID":"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5","Type":"ContainerStarted","Data":"223ee723725e75555f1a7c4ee9091dd3fe3c47c50e3cc0f69db38925595c0c90"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.428617 4809 generic.go:334] "Generic (PLEG): container finished" podID="3aede7af-428a-4d18-b78d-39e2c438d140" containerID="4692bcfe7142bc2738fd2b0178e41f14055eeab3c3ae7f26d1ad3477b9ee970c" exitCode=0 Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.428741 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jllxv" event={"ID":"3aede7af-428a-4d18-b78d-39e2c438d140","Type":"ContainerDied","Data":"4692bcfe7142bc2738fd2b0178e41f14055eeab3c3ae7f26d1ad3477b9ee970c"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.428785 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jllxv" event={"ID":"3aede7af-428a-4d18-b78d-39e2c438d140","Type":"ContainerStarted","Data":"361c7a7873dca2cd67e8fea3f6de6d309b94e9819b4354d13ee4fabb69d2f254"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.433874 4809 generic.go:334] "Generic (PLEG): container finished" podID="16be83eb-dd5e-4ddf-9d7d-ad6cba58a841" containerID="f395b9608a6f94d2a59890234bff67c7844f4946ecc0490217665ccdc20f313b" exitCode=0 Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.433976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lxhkk" event={"ID":"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841","Type":"ContainerDied","Data":"f395b9608a6f94d2a59890234bff67c7844f4946ecc0490217665ccdc20f313b"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.434027 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lxhkk" event={"ID":"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841","Type":"ContainerStarted","Data":"6f43e287c314165292a38be26511a5726af9657b4fe3639f8661dee256bbc88a"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.436326 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3aabcdcf-4f51-4398-ae40-33b6d57aa514","Type":"ContainerStarted","Data":"855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8"} Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.437093 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.438824 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-68v9p" podStartSLOduration=11.438813364 podStartE2EDuration="11.438813364s" podCreationTimestamp="2025-12-05 11:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:18.437477289 +0000 UTC m=+1253.828453847" watchObservedRunningTime="2025-12-05 11:29:18.438813364 +0000 UTC m=+1253.829789922" Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.510506 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=6.8508627650000005 podStartE2EDuration="13.510481099s" podCreationTimestamp="2025-12-05 11:29:05 +0000 UTC" firstStartedPulling="2025-12-05 11:29:06.451786615 +0000 UTC m=+1241.842763173" lastFinishedPulling="2025-12-05 11:29:13.111404949 +0000 UTC m=+1248.502381507" observedRunningTime="2025-12-05 11:29:18.491511203 +0000 UTC m=+1253.882487771" watchObservedRunningTime="2025-12-05 11:29:18.510481099 +0000 UTC m=+1253.901457657" Dec 05 11:29:18 crc kubenswrapper[4809]: I1205 11:29:18.543936 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-tcsq6" podStartSLOduration=2.659574702 podStartE2EDuration="9.543915954s" podCreationTimestamp="2025-12-05 11:29:09 +0000 UTC" firstStartedPulling="2025-12-05 11:29:10.319303232 +0000 UTC m=+1245.710279790" lastFinishedPulling="2025-12-05 11:29:17.203644474 +0000 UTC m=+1252.594621042" observedRunningTime="2025-12-05 11:29:18.538298067 +0000 UTC m=+1253.929274625" watchObservedRunningTime="2025-12-05 11:29:18.543915954 +0000 UTC m=+1253.934892502" Dec 05 11:29:19 crc kubenswrapper[4809]: I1205 11:29:19.587909 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 05 11:29:19 crc kubenswrapper[4809]: I1205 11:29:19.723543 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 05 11:29:19 crc kubenswrapper[4809]: I1205 11:29:19.878911 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7fd796d7df-gpkvt" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.051943 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.059268 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jllxv" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.071658 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.079094 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.084716 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aede7af-428a-4d18-b78d-39e2c438d140-operator-scripts\") pod \"3aede7af-428a-4d18-b78d-39e2c438d140\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.084810 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r58qw\" (UniqueName: \"kubernetes.io/projected/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-kube-api-access-r58qw\") pod \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.084849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m97td\" (UniqueName: \"kubernetes.io/projected/afcf740a-2a63-438c-8cd9-774324492a8e-kube-api-access-m97td\") pod \"afcf740a-2a63-438c-8cd9-774324492a8e\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.084908 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdtjd\" (UniqueName: \"kubernetes.io/projected/3aede7af-428a-4d18-b78d-39e2c438d140-kube-api-access-pdtjd\") pod \"3aede7af-428a-4d18-b78d-39e2c438d140\" (UID: \"3aede7af-428a-4d18-b78d-39e2c438d140\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.085223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-operator-scripts\") pod \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\" (UID: \"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.085250 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aede7af-428a-4d18-b78d-39e2c438d140-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3aede7af-428a-4d18-b78d-39e2c438d140" (UID: "3aede7af-428a-4d18-b78d-39e2c438d140"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.085315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcf740a-2a63-438c-8cd9-774324492a8e-operator-scripts\") pod \"afcf740a-2a63-438c-8cd9-774324492a8e\" (UID: \"afcf740a-2a63-438c-8cd9-774324492a8e\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.086006 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3aede7af-428a-4d18-b78d-39e2c438d140-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.086539 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16be83eb-dd5e-4ddf-9d7d-ad6cba58a841" (UID: "16be83eb-dd5e-4ddf-9d7d-ad6cba58a841"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.086601 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afcf740a-2a63-438c-8cd9-774324492a8e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afcf740a-2a63-438c-8cd9-774324492a8e" (UID: "afcf740a-2a63-438c-8cd9-774324492a8e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.093999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aede7af-428a-4d18-b78d-39e2c438d140-kube-api-access-pdtjd" (OuterVolumeSpecName: "kube-api-access-pdtjd") pod "3aede7af-428a-4d18-b78d-39e2c438d140" (UID: "3aede7af-428a-4d18-b78d-39e2c438d140"). InnerVolumeSpecName "kube-api-access-pdtjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.099776 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-kube-api-access-r58qw" (OuterVolumeSpecName: "kube-api-access-r58qw") pod "16be83eb-dd5e-4ddf-9d7d-ad6cba58a841" (UID: "16be83eb-dd5e-4ddf-9d7d-ad6cba58a841"). InnerVolumeSpecName "kube-api-access-r58qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.099879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afcf740a-2a63-438c-8cd9-774324492a8e-kube-api-access-m97td" (OuterVolumeSpecName: "kube-api-access-m97td") pod "afcf740a-2a63-438c-8cd9-774324492a8e" (UID: "afcf740a-2a63-438c-8cd9-774324492a8e"). InnerVolumeSpecName "kube-api-access-m97td". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187071 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-operator-scripts\") pod \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22xnz\" (UniqueName: \"kubernetes.io/projected/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-kube-api-access-22xnz\") pod \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\" (UID: \"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5\") " Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187501 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d074606c-4b4c-46bc-8552-d9a1ddcd9cc5" (UID: "d074606c-4b4c-46bc-8552-d9a1ddcd9cc5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187835 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r58qw\" (UniqueName: \"kubernetes.io/projected/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-kube-api-access-r58qw\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187851 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m97td\" (UniqueName: \"kubernetes.io/projected/afcf740a-2a63-438c-8cd9-774324492a8e-kube-api-access-m97td\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187861 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdtjd\" (UniqueName: \"kubernetes.io/projected/3aede7af-428a-4d18-b78d-39e2c438d140-kube-api-access-pdtjd\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187872 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187881 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcf740a-2a63-438c-8cd9-774324492a8e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.187889 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.190274 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-kube-api-access-22xnz" (OuterVolumeSpecName: "kube-api-access-22xnz") pod "d074606c-4b4c-46bc-8552-d9a1ddcd9cc5" (UID: "d074606c-4b4c-46bc-8552-d9a1ddcd9cc5"). InnerVolumeSpecName "kube-api-access-22xnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.290099 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22xnz\" (UniqueName: \"kubernetes.io/projected/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5-kube-api-access-22xnz\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.397974 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.489022 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-838e-account-create-update-rt9bd" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.489022 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-838e-account-create-update-rt9bd" event={"ID":"afcf740a-2a63-438c-8cd9-774324492a8e","Type":"ContainerDied","Data":"d7dc30602b6526c2c13a5a47b8192b5defa06b965d3fdc1028fca43e5c70cf88"} Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.489091 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7dc30602b6526c2c13a5a47b8192b5defa06b965d3fdc1028fca43e5c70cf88" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.491264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ef3a-account-create-update-9mdwj" event={"ID":"d074606c-4b4c-46bc-8552-d9a1ddcd9cc5","Type":"ContainerDied","Data":"223ee723725e75555f1a7c4ee9091dd3fe3c47c50e3cc0f69db38925595c0c90"} Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.491298 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="223ee723725e75555f1a7c4ee9091dd3fe3c47c50e3cc0f69db38925595c0c90" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.491370 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ef3a-account-create-update-9mdwj" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.501711 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jllxv" event={"ID":"3aede7af-428a-4d18-b78d-39e2c438d140","Type":"ContainerDied","Data":"361c7a7873dca2cd67e8fea3f6de6d309b94e9819b4354d13ee4fabb69d2f254"} Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.501785 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="361c7a7873dca2cd67e8fea3f6de6d309b94e9819b4354d13ee4fabb69d2f254" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.501917 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jllxv" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.508580 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lxhkk" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.509502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lxhkk" event={"ID":"16be83eb-dd5e-4ddf-9d7d-ad6cba58a841","Type":"ContainerDied","Data":"6f43e287c314165292a38be26511a5726af9657b4fe3639f8661dee256bbc88a"} Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.509537 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f43e287c314165292a38be26511a5726af9657b4fe3639f8661dee256bbc88a" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.971535 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-4ng99"] Dec 05 11:29:20 crc kubenswrapper[4809]: E1205 11:29:20.972240 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16be83eb-dd5e-4ddf-9d7d-ad6cba58a841" containerName="mariadb-database-create" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972255 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="16be83eb-dd5e-4ddf-9d7d-ad6cba58a841" containerName="mariadb-database-create" Dec 05 11:29:20 crc kubenswrapper[4809]: E1205 11:29:20.972269 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aede7af-428a-4d18-b78d-39e2c438d140" containerName="mariadb-database-create" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972275 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aede7af-428a-4d18-b78d-39e2c438d140" containerName="mariadb-database-create" Dec 05 11:29:20 crc kubenswrapper[4809]: E1205 11:29:20.972292 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afcf740a-2a63-438c-8cd9-774324492a8e" containerName="mariadb-account-create-update" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972298 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="afcf740a-2a63-438c-8cd9-774324492a8e" containerName="mariadb-account-create-update" Dec 05 11:29:20 crc kubenswrapper[4809]: E1205 11:29:20.972307 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" containerName="dnsmasq-dns" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972313 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" containerName="dnsmasq-dns" Dec 05 11:29:20 crc kubenswrapper[4809]: E1205 11:29:20.972323 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d074606c-4b4c-46bc-8552-d9a1ddcd9cc5" containerName="mariadb-account-create-update" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972329 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d074606c-4b4c-46bc-8552-d9a1ddcd9cc5" containerName="mariadb-account-create-update" Dec 05 11:29:20 crc kubenswrapper[4809]: E1205 11:29:20.972343 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" containerName="init" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972348 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" containerName="init" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972494 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d074606c-4b4c-46bc-8552-d9a1ddcd9cc5" containerName="mariadb-account-create-update" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972518 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae470489-d955-443f-a16b-aa09f3955cd2" containerName="dnsmasq-dns" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972542 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aede7af-428a-4d18-b78d-39e2c438d140" containerName="mariadb-database-create" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972558 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="16be83eb-dd5e-4ddf-9d7d-ad6cba58a841" containerName="mariadb-database-create" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.972574 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="afcf740a-2a63-438c-8cd9-774324492a8e" containerName="mariadb-account-create-update" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.973195 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4ng99" Dec 05 11:29:20 crc kubenswrapper[4809]: I1205 11:29:20.985983 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4ng99"] Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.004803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced4cdcc-ec40-4177-a706-0386c12f1529-operator-scripts\") pod \"glance-db-create-4ng99\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " pod="openstack/glance-db-create-4ng99" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.004980 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-829zd\" (UniqueName: \"kubernetes.io/projected/ced4cdcc-ec40-4177-a706-0386c12f1529-kube-api-access-829zd\") pod \"glance-db-create-4ng99\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " pod="openstack/glance-db-create-4ng99" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.072241 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-21cd-account-create-update-hblhl"] Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.073833 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.082524 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.083868 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-21cd-account-create-update-hblhl"] Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.107315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z9rf\" (UniqueName: \"kubernetes.io/projected/c82cff53-0da9-4bed-baf0-4aa69f5c9074-kube-api-access-6z9rf\") pod \"glance-21cd-account-create-update-hblhl\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.107456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced4cdcc-ec40-4177-a706-0386c12f1529-operator-scripts\") pod \"glance-db-create-4ng99\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " pod="openstack/glance-db-create-4ng99" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.107744 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-829zd\" (UniqueName: \"kubernetes.io/projected/ced4cdcc-ec40-4177-a706-0386c12f1529-kube-api-access-829zd\") pod \"glance-db-create-4ng99\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " pod="openstack/glance-db-create-4ng99" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.107829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c82cff53-0da9-4bed-baf0-4aa69f5c9074-operator-scripts\") pod \"glance-21cd-account-create-update-hblhl\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.108371 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced4cdcc-ec40-4177-a706-0386c12f1529-operator-scripts\") pod \"glance-db-create-4ng99\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " pod="openstack/glance-db-create-4ng99" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.126438 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-829zd\" (UniqueName: \"kubernetes.io/projected/ced4cdcc-ec40-4177-a706-0386c12f1529-kube-api-access-829zd\") pod \"glance-db-create-4ng99\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " pod="openstack/glance-db-create-4ng99" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.209693 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c82cff53-0da9-4bed-baf0-4aa69f5c9074-operator-scripts\") pod \"glance-21cd-account-create-update-hblhl\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.210031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z9rf\" (UniqueName: \"kubernetes.io/projected/c82cff53-0da9-4bed-baf0-4aa69f5c9074-kube-api-access-6z9rf\") pod \"glance-21cd-account-create-update-hblhl\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.211230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c82cff53-0da9-4bed-baf0-4aa69f5c9074-operator-scripts\") pod \"glance-21cd-account-create-update-hblhl\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.226198 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z9rf\" (UniqueName: \"kubernetes.io/projected/c82cff53-0da9-4bed-baf0-4aa69f5c9074-kube-api-access-6z9rf\") pod \"glance-21cd-account-create-update-hblhl\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.291580 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4ng99" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.393095 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.724136 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4ng99"] Dec 05 11:29:21 crc kubenswrapper[4809]: W1205 11:29:21.725464 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podced4cdcc_ec40_4177_a706_0386c12f1529.slice/crio-459e37840ef03e858afd6ae32100e3f366e42e2a34e1064437c70e1f77636d62 WatchSource:0}: Error finding container 459e37840ef03e858afd6ae32100e3f366e42e2a34e1064437c70e1f77636d62: Status 404 returned error can't find the container with id 459e37840ef03e858afd6ae32100e3f366e42e2a34e1064437c70e1f77636d62 Dec 05 11:29:21 crc kubenswrapper[4809]: I1205 11:29:21.828021 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-21cd-account-create-update-hblhl"] Dec 05 11:29:21 crc kubenswrapper[4809]: W1205 11:29:21.833778 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc82cff53_0da9_4bed_baf0_4aa69f5c9074.slice/crio-8d0f3eaf41dbbce8124f94b7ce2a44a68f229b88c471ab4959efb402d9572cf5 WatchSource:0}: Error finding container 8d0f3eaf41dbbce8124f94b7ce2a44a68f229b88c471ab4959efb402d9572cf5: Status 404 returned error can't find the container with id 8d0f3eaf41dbbce8124f94b7ce2a44a68f229b88c471ab4959efb402d9572cf5 Dec 05 11:29:22 crc kubenswrapper[4809]: I1205 11:29:22.523928 4809 generic.go:334] "Generic (PLEG): container finished" podID="c82cff53-0da9-4bed-baf0-4aa69f5c9074" containerID="ab993a42330cf1af383e0abbeeb94faa4345af10b7cf650185f7e021bb2a7d93" exitCode=0 Dec 05 11:29:22 crc kubenswrapper[4809]: I1205 11:29:22.523987 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-21cd-account-create-update-hblhl" event={"ID":"c82cff53-0da9-4bed-baf0-4aa69f5c9074","Type":"ContainerDied","Data":"ab993a42330cf1af383e0abbeeb94faa4345af10b7cf650185f7e021bb2a7d93"} Dec 05 11:29:22 crc kubenswrapper[4809]: I1205 11:29:22.524308 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-21cd-account-create-update-hblhl" event={"ID":"c82cff53-0da9-4bed-baf0-4aa69f5c9074","Type":"ContainerStarted","Data":"8d0f3eaf41dbbce8124f94b7ce2a44a68f229b88c471ab4959efb402d9572cf5"} Dec 05 11:29:22 crc kubenswrapper[4809]: I1205 11:29:22.525736 4809 generic.go:334] "Generic (PLEG): container finished" podID="ced4cdcc-ec40-4177-a706-0386c12f1529" containerID="b5512e72b384254d31928502d46c73cced07ecb562b80bd83398a34c2cdbb8ef" exitCode=0 Dec 05 11:29:22 crc kubenswrapper[4809]: I1205 11:29:22.525771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4ng99" event={"ID":"ced4cdcc-ec40-4177-a706-0386c12f1529","Type":"ContainerDied","Data":"b5512e72b384254d31928502d46c73cced07ecb562b80bd83398a34c2cdbb8ef"} Dec 05 11:29:22 crc kubenswrapper[4809]: I1205 11:29:22.525792 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4ng99" event={"ID":"ced4cdcc-ec40-4177-a706-0386c12f1529","Type":"ContainerStarted","Data":"459e37840ef03e858afd6ae32100e3f366e42e2a34e1064437c70e1f77636d62"} Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.236785 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.296006 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8shnb"] Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.296562 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" podUID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerName="dnsmasq-dns" containerID="cri-o://191de9b1fc5f9c1b4fce32c5d2c9973de7db04c83f0f2a9a384bbd4821136087" gracePeriod=10 Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.535949 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerID="191de9b1fc5f9c1b4fce32c5d2c9973de7db04c83f0f2a9a384bbd4821136087" exitCode=0 Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.536183 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" event={"ID":"7d49f737-1a41-4f92-b47e-d2935151a71c","Type":"ContainerDied","Data":"191de9b1fc5f9c1b4fce32c5d2c9973de7db04c83f0f2a9a384bbd4821136087"} Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.789721 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.941408 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.954008 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-config\") pod \"7d49f737-1a41-4f92-b47e-d2935151a71c\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.954148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-nb\") pod \"7d49f737-1a41-4f92-b47e-d2935151a71c\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.954174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-dns-svc\") pod \"7d49f737-1a41-4f92-b47e-d2935151a71c\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.954194 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-sb\") pod \"7d49f737-1a41-4f92-b47e-d2935151a71c\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.954271 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vddz\" (UniqueName: \"kubernetes.io/projected/7d49f737-1a41-4f92-b47e-d2935151a71c-kube-api-access-2vddz\") pod \"7d49f737-1a41-4f92-b47e-d2935151a71c\" (UID: \"7d49f737-1a41-4f92-b47e-d2935151a71c\") " Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.955148 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4ng99" Dec 05 11:29:23 crc kubenswrapper[4809]: I1205 11:29:23.961269 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d49f737-1a41-4f92-b47e-d2935151a71c-kube-api-access-2vddz" (OuterVolumeSpecName: "kube-api-access-2vddz") pod "7d49f737-1a41-4f92-b47e-d2935151a71c" (UID: "7d49f737-1a41-4f92-b47e-d2935151a71c"). InnerVolumeSpecName "kube-api-access-2vddz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.003141 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-config" (OuterVolumeSpecName: "config") pod "7d49f737-1a41-4f92-b47e-d2935151a71c" (UID: "7d49f737-1a41-4f92-b47e-d2935151a71c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.006040 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d49f737-1a41-4f92-b47e-d2935151a71c" (UID: "7d49f737-1a41-4f92-b47e-d2935151a71c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.018140 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7d49f737-1a41-4f92-b47e-d2935151a71c" (UID: "7d49f737-1a41-4f92-b47e-d2935151a71c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.032715 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7d49f737-1a41-4f92-b47e-d2935151a71c" (UID: "7d49f737-1a41-4f92-b47e-d2935151a71c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.055992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced4cdcc-ec40-4177-a706-0386c12f1529-operator-scripts\") pod \"ced4cdcc-ec40-4177-a706-0386c12f1529\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056096 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-829zd\" (UniqueName: \"kubernetes.io/projected/ced4cdcc-ec40-4177-a706-0386c12f1529-kube-api-access-829zd\") pod \"ced4cdcc-ec40-4177-a706-0386c12f1529\" (UID: \"ced4cdcc-ec40-4177-a706-0386c12f1529\") " Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z9rf\" (UniqueName: \"kubernetes.io/projected/c82cff53-0da9-4bed-baf0-4aa69f5c9074-kube-api-access-6z9rf\") pod \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056194 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c82cff53-0da9-4bed-baf0-4aa69f5c9074-operator-scripts\") pod \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\" (UID: \"c82cff53-0da9-4bed-baf0-4aa69f5c9074\") " Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056605 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vddz\" (UniqueName: \"kubernetes.io/projected/7d49f737-1a41-4f92-b47e-d2935151a71c-kube-api-access-2vddz\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056624 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056643 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056652 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.056660 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d49f737-1a41-4f92-b47e-d2935151a71c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.057924 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced4cdcc-ec40-4177-a706-0386c12f1529-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ced4cdcc-ec40-4177-a706-0386c12f1529" (UID: "ced4cdcc-ec40-4177-a706-0386c12f1529"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.058496 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c82cff53-0da9-4bed-baf0-4aa69f5c9074-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c82cff53-0da9-4bed-baf0-4aa69f5c9074" (UID: "c82cff53-0da9-4bed-baf0-4aa69f5c9074"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.059785 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced4cdcc-ec40-4177-a706-0386c12f1529-kube-api-access-829zd" (OuterVolumeSpecName: "kube-api-access-829zd") pod "ced4cdcc-ec40-4177-a706-0386c12f1529" (UID: "ced4cdcc-ec40-4177-a706-0386c12f1529"). InnerVolumeSpecName "kube-api-access-829zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.060287 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c82cff53-0da9-4bed-baf0-4aa69f5c9074-kube-api-access-6z9rf" (OuterVolumeSpecName: "kube-api-access-6z9rf") pod "c82cff53-0da9-4bed-baf0-4aa69f5c9074" (UID: "c82cff53-0da9-4bed-baf0-4aa69f5c9074"). InnerVolumeSpecName "kube-api-access-6z9rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.158805 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced4cdcc-ec40-4177-a706-0386c12f1529-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.158859 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-829zd\" (UniqueName: \"kubernetes.io/projected/ced4cdcc-ec40-4177-a706-0386c12f1529-kube-api-access-829zd\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.158882 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z9rf\" (UniqueName: \"kubernetes.io/projected/c82cff53-0da9-4bed-baf0-4aa69f5c9074-kube-api-access-6z9rf\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.158902 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c82cff53-0da9-4bed-baf0-4aa69f5c9074-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.544725 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-21cd-account-create-update-hblhl" event={"ID":"c82cff53-0da9-4bed-baf0-4aa69f5c9074","Type":"ContainerDied","Data":"8d0f3eaf41dbbce8124f94b7ce2a44a68f229b88c471ab4959efb402d9572cf5"} Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.544784 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d0f3eaf41dbbce8124f94b7ce2a44a68f229b88c471ab4959efb402d9572cf5" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.544748 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-21cd-account-create-update-hblhl" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.546594 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4ng99" event={"ID":"ced4cdcc-ec40-4177-a706-0386c12f1529","Type":"ContainerDied","Data":"459e37840ef03e858afd6ae32100e3f366e42e2a34e1064437c70e1f77636d62"} Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.546724 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="459e37840ef03e858afd6ae32100e3f366e42e2a34e1064437c70e1f77636d62" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.546597 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4ng99" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.548771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" event={"ID":"7d49f737-1a41-4f92-b47e-d2935151a71c","Type":"ContainerDied","Data":"8add03e920e8ac974244bf14db514f9bb082e851dda7c82c8d22ef71334aeeb6"} Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.548821 4809 scope.go:117] "RemoveContainer" containerID="191de9b1fc5f9c1b4fce32c5d2c9973de7db04c83f0f2a9a384bbd4821136087" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.548833 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8shnb" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.576382 4809 scope.go:117] "RemoveContainer" containerID="b189039130105de78915b07ae0163f1ed73b92a55d08d33e2692e02c7760b398" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.602593 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8shnb"] Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.609272 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8shnb"] Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.879622 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.887478 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"swift-storage-0\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " pod="openstack/swift-storage-0" Dec 05 11:29:24 crc kubenswrapper[4809]: I1205 11:29:24.897954 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d49f737-1a41-4f92-b47e-d2935151a71c" path="/var/lib/kubelet/pods/7d49f737-1a41-4f92-b47e-d2935151a71c/volumes" Dec 05 11:29:25 crc kubenswrapper[4809]: I1205 11:29:25.185175 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 11:29:25 crc kubenswrapper[4809]: I1205 11:29:25.559682 4809 generic.go:334] "Generic (PLEG): container finished" podID="ac85b554-c4a8-483f-a64d-cedf0984140e" containerID="bebae10222675f498e6edbdca8558f09457b6b61eba7576c606fad86d41f948e" exitCode=0 Dec 05 11:29:25 crc kubenswrapper[4809]: I1205 11:29:25.559785 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcsq6" event={"ID":"ac85b554-c4a8-483f-a64d-cedf0984140e","Type":"ContainerDied","Data":"bebae10222675f498e6edbdca8558f09457b6b61eba7576c606fad86d41f948e"} Dec 05 11:29:25 crc kubenswrapper[4809]: I1205 11:29:25.781737 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:29:25 crc kubenswrapper[4809]: W1205 11:29:25.788787 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58501ab1_18d7_45d9_98cb_a74e62db36fc.slice/crio-ebdbfd0085c5a8b87e7a702a3c0c3d81bf14bb2eec0523014bd68a71b4c5e402 WatchSource:0}: Error finding container ebdbfd0085c5a8b87e7a702a3c0c3d81bf14bb2eec0523014bd68a71b4c5e402: Status 404 returned error can't find the container with id ebdbfd0085c5a8b87e7a702a3c0c3d81bf14bb2eec0523014bd68a71b4c5e402 Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.070030 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.335432 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-5mwp9"] Dec 05 11:29:26 crc kubenswrapper[4809]: E1205 11:29:26.335906 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerName="init" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.335929 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerName="init" Dec 05 11:29:26 crc kubenswrapper[4809]: E1205 11:29:26.335950 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerName="dnsmasq-dns" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.335959 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerName="dnsmasq-dns" Dec 05 11:29:26 crc kubenswrapper[4809]: E1205 11:29:26.335972 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced4cdcc-ec40-4177-a706-0386c12f1529" containerName="mariadb-database-create" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.335980 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced4cdcc-ec40-4177-a706-0386c12f1529" containerName="mariadb-database-create" Dec 05 11:29:26 crc kubenswrapper[4809]: E1205 11:29:26.336000 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c82cff53-0da9-4bed-baf0-4aa69f5c9074" containerName="mariadb-account-create-update" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.336008 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c82cff53-0da9-4bed-baf0-4aa69f5c9074" containerName="mariadb-account-create-update" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.336209 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c82cff53-0da9-4bed-baf0-4aa69f5c9074" containerName="mariadb-account-create-update" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.336245 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d49f737-1a41-4f92-b47e-d2935151a71c" containerName="dnsmasq-dns" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.336274 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced4cdcc-ec40-4177-a706-0386c12f1529" containerName="mariadb-database-create" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.336975 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.347798 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.355387 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5mwp9"] Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.358100 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hfvxz" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.402089 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw4bm\" (UniqueName: \"kubernetes.io/projected/b4e08826-2c05-4f4f-b899-7a87f44e51d5-kube-api-access-nw4bm\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.402388 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-combined-ca-bundle\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.402531 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-config-data\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.402693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-db-sync-config-data\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.503878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-db-sync-config-data\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.503942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw4bm\" (UniqueName: \"kubernetes.io/projected/b4e08826-2c05-4f4f-b899-7a87f44e51d5-kube-api-access-nw4bm\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.504010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-combined-ca-bundle\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.504085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-config-data\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.515373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-db-sync-config-data\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.518950 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-combined-ca-bundle\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.519785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-config-data\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.520237 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw4bm\" (UniqueName: \"kubernetes.io/projected/b4e08826-2c05-4f4f-b899-7a87f44e51d5-kube-api-access-nw4bm\") pod \"glance-db-sync-5mwp9\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.577738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"ebdbfd0085c5a8b87e7a702a3c0c3d81bf14bb2eec0523014bd68a71b4c5e402"} Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.668045 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:26 crc kubenswrapper[4809]: I1205 11:29:26.960799 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.114523 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-scripts\") pod \"ac85b554-c4a8-483f-a64d-cedf0984140e\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.114648 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac85b554-c4a8-483f-a64d-cedf0984140e-etc-swift\") pod \"ac85b554-c4a8-483f-a64d-cedf0984140e\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.114696 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-ring-data-devices\") pod \"ac85b554-c4a8-483f-a64d-cedf0984140e\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.114723 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-dispersionconf\") pod \"ac85b554-c4a8-483f-a64d-cedf0984140e\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.114786 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ml6t\" (UniqueName: \"kubernetes.io/projected/ac85b554-c4a8-483f-a64d-cedf0984140e-kube-api-access-5ml6t\") pod \"ac85b554-c4a8-483f-a64d-cedf0984140e\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.114839 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-combined-ca-bundle\") pod \"ac85b554-c4a8-483f-a64d-cedf0984140e\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.114864 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-swiftconf\") pod \"ac85b554-c4a8-483f-a64d-cedf0984140e\" (UID: \"ac85b554-c4a8-483f-a64d-cedf0984140e\") " Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.116553 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ac85b554-c4a8-483f-a64d-cedf0984140e" (UID: "ac85b554-c4a8-483f-a64d-cedf0984140e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.118441 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac85b554-c4a8-483f-a64d-cedf0984140e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ac85b554-c4a8-483f-a64d-cedf0984140e" (UID: "ac85b554-c4a8-483f-a64d-cedf0984140e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.124113 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac85b554-c4a8-483f-a64d-cedf0984140e-kube-api-access-5ml6t" (OuterVolumeSpecName: "kube-api-access-5ml6t") pod "ac85b554-c4a8-483f-a64d-cedf0984140e" (UID: "ac85b554-c4a8-483f-a64d-cedf0984140e"). InnerVolumeSpecName "kube-api-access-5ml6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.131089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ac85b554-c4a8-483f-a64d-cedf0984140e" (UID: "ac85b554-c4a8-483f-a64d-cedf0984140e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.146532 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ac85b554-c4a8-483f-a64d-cedf0984140e" (UID: "ac85b554-c4a8-483f-a64d-cedf0984140e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.148090 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac85b554-c4a8-483f-a64d-cedf0984140e" (UID: "ac85b554-c4a8-483f-a64d-cedf0984140e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.149266 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-scripts" (OuterVolumeSpecName: "scripts") pod "ac85b554-c4a8-483f-a64d-cedf0984140e" (UID: "ac85b554-c4a8-483f-a64d-cedf0984140e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.216224 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ml6t\" (UniqueName: \"kubernetes.io/projected/ac85b554-c4a8-483f-a64d-cedf0984140e-kube-api-access-5ml6t\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.216575 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.216588 4809 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.216598 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.216607 4809 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ac85b554-c4a8-483f-a64d-cedf0984140e-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.216615 4809 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ac85b554-c4a8-483f-a64d-cedf0984140e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.216623 4809 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ac85b554-c4a8-483f-a64d-cedf0984140e-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.477956 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5mwp9"] Dec 05 11:29:27 crc kubenswrapper[4809]: W1205 11:29:27.482581 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4e08826_2c05_4f4f_b899_7a87f44e51d5.slice/crio-485da7f3ff051ba72459adb79eb97be8276176e361305c2af90c641725b0ff71 WatchSource:0}: Error finding container 485da7f3ff051ba72459adb79eb97be8276176e361305c2af90c641725b0ff71: Status 404 returned error can't find the container with id 485da7f3ff051ba72459adb79eb97be8276176e361305c2af90c641725b0ff71 Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.590685 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5mwp9" event={"ID":"b4e08826-2c05-4f4f-b899-7a87f44e51d5","Type":"ContainerStarted","Data":"485da7f3ff051ba72459adb79eb97be8276176e361305c2af90c641725b0ff71"} Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.592904 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcsq6" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.592991 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcsq6" event={"ID":"ac85b554-c4a8-483f-a64d-cedf0984140e","Type":"ContainerDied","Data":"0acd9c9e86a3931b37f4284e1c2feffafdcdc7206669e36d03e2b7b44a9c25dd"} Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.593046 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0acd9c9e86a3931b37f4284e1c2feffafdcdc7206669e36d03e2b7b44a9c25dd" Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.596484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"0bb5aa4d3d56722e820bbab3a1408f0e2a8ad15ee9a7ade8db2984056113caab"} Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.596537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"2af085bac46d211c6f1357baf53c68b8c882c87ee5a214e91c8e24cfb34aaa57"} Dec 05 11:29:27 crc kubenswrapper[4809]: I1205 11:29:27.596553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"5cdc142e3c8a66238160b2d52e5885ba4ddb6262d5ad5802d543fa708c536590"} Dec 05 11:29:28 crc kubenswrapper[4809]: I1205 11:29:28.609661 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"c952b386502163d911d4615800b364440abeafbe8af4d4762f426ad2d8920b4c"} Dec 05 11:29:29 crc kubenswrapper[4809]: I1205 11:29:29.622864 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"d80d9b0fd1b858fe84af319a01ca8ae68a9151b6c070e69dff8050d858ebc3be"} Dec 05 11:29:29 crc kubenswrapper[4809]: I1205 11:29:29.623267 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"58ee5b8b26614f934ab1ca4e41f1129f8f4b37fb56c414f5eef99f29744b4275"} Dec 05 11:29:30 crc kubenswrapper[4809]: I1205 11:29:30.636058 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"213d6625154c146c9050604a6a30c02a271e04a3445369142a80d5fbd10042c8"} Dec 05 11:29:30 crc kubenswrapper[4809]: I1205 11:29:30.636111 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"f644863166060fc77243c2059ca8902c14f11a6c9b85f76eb1d36cd09decb692"} Dec 05 11:29:31 crc kubenswrapper[4809]: I1205 11:29:31.648175 4809 generic.go:334] "Generic (PLEG): container finished" podID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerID="03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1" exitCode=0 Dec 05 11:29:31 crc kubenswrapper[4809]: I1205 11:29:31.648257 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e1251531-cb99-418e-aa38-65a793e5a9d0","Type":"ContainerDied","Data":"03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1"} Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.254264 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fg4wz" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" probeResult="failure" output=< Dec 05 11:29:32 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 11:29:32 crc kubenswrapper[4809]: > Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.265465 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.265518 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.491877 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fg4wz-config-q7bdj"] Dec 05 11:29:32 crc kubenswrapper[4809]: E1205 11:29:32.493590 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac85b554-c4a8-483f-a64d-cedf0984140e" containerName="swift-ring-rebalance" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.493615 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac85b554-c4a8-483f-a64d-cedf0984140e" containerName="swift-ring-rebalance" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.493832 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac85b554-c4a8-483f-a64d-cedf0984140e" containerName="swift-ring-rebalance" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.494461 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.496284 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.506745 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fg4wz-config-q7bdj"] Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.612836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run-ovn\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.612916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-log-ovn\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.613066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-scripts\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.613107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgz9v\" (UniqueName: \"kubernetes.io/projected/c8ed3bb8-646b-4db7-9b7f-556f5625d104-kube-api-access-vgz9v\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.613130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.613162 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-additional-scripts\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.664551 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"6bb7e4f8fe910073653aa1a9f3a5a433f92b9f57a52d341885532226825ec2ca"} Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.664595 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"fc97466f1d8534b47cb24e449245a63275f1660daf521b08f83aaabeed242660"} Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.666468 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e1251531-cb99-418e-aa38-65a793e5a9d0","Type":"ContainerStarted","Data":"aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30"} Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.668037 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.698509 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.73671155 podStartE2EDuration="1m22.698489987s" podCreationTimestamp="2025-12-05 11:28:10 +0000 UTC" firstStartedPulling="2025-12-05 11:28:12.4607988 +0000 UTC m=+1187.851775358" lastFinishedPulling="2025-12-05 11:28:58.422577237 +0000 UTC m=+1233.813553795" observedRunningTime="2025-12-05 11:29:32.689703577 +0000 UTC m=+1268.080680155" watchObservedRunningTime="2025-12-05 11:29:32.698489987 +0000 UTC m=+1268.089466555" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714125 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgz9v\" (UniqueName: \"kubernetes.io/projected/c8ed3bb8-646b-4db7-9b7f-556f5625d104-kube-api-access-vgz9v\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714233 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-additional-scripts\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run-ovn\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714317 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-log-ovn\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-scripts\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714651 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.714865 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run-ovn\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.715544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-additional-scripts\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.715654 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-log-ovn\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.717132 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-scripts\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.733386 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgz9v\" (UniqueName: \"kubernetes.io/projected/c8ed3bb8-646b-4db7-9b7f-556f5625d104-kube-api-access-vgz9v\") pod \"ovn-controller-fg4wz-config-q7bdj\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:32 crc kubenswrapper[4809]: I1205 11:29:32.821980 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:37 crc kubenswrapper[4809]: I1205 11:29:37.251828 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fg4wz" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" probeResult="failure" output=< Dec 05 11:29:37 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 11:29:37 crc kubenswrapper[4809]: > Dec 05 11:29:37 crc kubenswrapper[4809]: I1205 11:29:37.706988 4809 generic.go:334] "Generic (PLEG): container finished" podID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerID="406129ebd40a9895b56eb4df3cb21847834b08c83a4aab01708e82905b319f2a" exitCode=0 Dec 05 11:29:37 crc kubenswrapper[4809]: I1205 11:29:37.707029 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfe522ab-449f-433e-a9bf-b33ed3b42595","Type":"ContainerDied","Data":"406129ebd40a9895b56eb4df3cb21847834b08c83a4aab01708e82905b319f2a"} Dec 05 11:29:39 crc kubenswrapper[4809]: I1205 11:29:39.374583 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fg4wz-config-q7bdj"] Dec 05 11:29:39 crc kubenswrapper[4809]: W1205 11:29:39.385486 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8ed3bb8_646b_4db7_9b7f_556f5625d104.slice/crio-d6f7280bb0e86d631b4856387bc3296bafbaad3a6fd983b0a0f44724887cbe3c WatchSource:0}: Error finding container d6f7280bb0e86d631b4856387bc3296bafbaad3a6fd983b0a0f44724887cbe3c: Status 404 returned error can't find the container with id d6f7280bb0e86d631b4856387bc3296bafbaad3a6fd983b0a0f44724887cbe3c Dec 05 11:29:39 crc kubenswrapper[4809]: I1205 11:29:39.725465 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz-config-q7bdj" event={"ID":"c8ed3bb8-646b-4db7-9b7f-556f5625d104","Type":"ContainerStarted","Data":"d6f7280bb0e86d631b4856387bc3296bafbaad3a6fd983b0a0f44724887cbe3c"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.737425 4809 generic.go:334] "Generic (PLEG): container finished" podID="c8ed3bb8-646b-4db7-9b7f-556f5625d104" containerID="a81cb4726f7d91229efe101694be71b159de9cb7acbee23af76ed162d4c42eab" exitCode=0 Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.737791 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz-config-q7bdj" event={"ID":"c8ed3bb8-646b-4db7-9b7f-556f5625d104","Type":"ContainerDied","Data":"a81cb4726f7d91229efe101694be71b159de9cb7acbee23af76ed162d4c42eab"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.740487 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfe522ab-449f-433e-a9bf-b33ed3b42595","Type":"ContainerStarted","Data":"b1d76863d0e54906522b49358c89ce129152fe39aef470b4c3740799e8e8769d"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.746274 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.747842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5mwp9" event={"ID":"b4e08826-2c05-4f4f-b899-7a87f44e51d5","Type":"ContainerStarted","Data":"32e6b77cca37dcdbc635d76201fe078ccada3173763f886d6471946dd5311728"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.771733 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"e13456d6c296dda4598f34ee0b8c0e4703c954f9a8553fd24476fb75c023f44c"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.771782 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"9d5533e9d925f25fad5f31fbd6f5b9fe99bee0d29f635ea849b663a235a46895"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.771803 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"640615ea9e04efceac48bfd84c7eaa5885da503591e80a899131fe2037fed294"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.771815 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"22c2f8cd7c22d4a0109b044a886f55e278f8c205daf14ca5e1b47bd41b23426a"} Dec 05 11:29:40 crc kubenswrapper[4809]: I1205 11:29:40.819955 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371947.03484 podStartE2EDuration="1m29.819936701s" podCreationTimestamp="2025-12-05 11:28:11 +0000 UTC" firstStartedPulling="2025-12-05 11:28:13.425708067 +0000 UTC m=+1188.816684625" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:40.799446965 +0000 UTC m=+1276.190423533" watchObservedRunningTime="2025-12-05 11:29:40.819936701 +0000 UTC m=+1276.210913259" Dec 05 11:29:41 crc kubenswrapper[4809]: I1205 11:29:41.786492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerStarted","Data":"0414b93f70cfeebd7e27ab579af01c51e47ad9419cbb11d16f41369b2fe5cc96"} Dec 05 11:29:41 crc kubenswrapper[4809]: I1205 11:29:41.830166 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=28.952360728 podStartE2EDuration="34.830146914s" podCreationTimestamp="2025-12-05 11:29:07 +0000 UTC" firstStartedPulling="2025-12-05 11:29:25.790588507 +0000 UTC m=+1261.181565085" lastFinishedPulling="2025-12-05 11:29:31.668374713 +0000 UTC m=+1267.059351271" observedRunningTime="2025-12-05 11:29:41.819460895 +0000 UTC m=+1277.210437463" watchObservedRunningTime="2025-12-05 11:29:41.830146914 +0000 UTC m=+1277.221123472" Dec 05 11:29:41 crc kubenswrapper[4809]: I1205 11:29:41.831711 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-5mwp9" podStartSLOduration=4.253336439 podStartE2EDuration="15.831704635s" podCreationTimestamp="2025-12-05 11:29:26 +0000 UTC" firstStartedPulling="2025-12-05 11:29:27.485502306 +0000 UTC m=+1262.876478874" lastFinishedPulling="2025-12-05 11:29:39.063870492 +0000 UTC m=+1274.454847070" observedRunningTime="2025-12-05 11:29:40.815071154 +0000 UTC m=+1276.206047722" watchObservedRunningTime="2025-12-05 11:29:41.831704635 +0000 UTC m=+1277.222681193" Dec 05 11:29:41 crc kubenswrapper[4809]: I1205 11:29:41.850095 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.158201 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-zmh58"] Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.159733 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.173761 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.185878 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-zmh58"] Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.189317 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.253995 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-fg4wz" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.288780 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-scripts\") pod \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.288849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-log-ovn\") pod \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.288881 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run-ovn\") pod \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.288953 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run\") pod \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289069 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgz9v\" (UniqueName: \"kubernetes.io/projected/c8ed3bb8-646b-4db7-9b7f-556f5625d104-kube-api-access-vgz9v\") pod \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-additional-scripts\") pod \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\" (UID: \"c8ed3bb8-646b-4db7-9b7f-556f5625d104\") " Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c8ed3bb8-646b-4db7-9b7f-556f5625d104" (UID: "c8ed3bb8-646b-4db7-9b7f-556f5625d104"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run" (OuterVolumeSpecName: "var-run") pod "c8ed3bb8-646b-4db7-9b7f-556f5625d104" (UID: "c8ed3bb8-646b-4db7-9b7f-556f5625d104"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289367 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289360 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c8ed3bb8-646b-4db7-9b7f-556f5625d104" (UID: "c8ed3bb8-646b-4db7-9b7f-556f5625d104"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289456 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9t8j\" (UniqueName: \"kubernetes.io/projected/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-kube-api-access-f9t8j\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-config\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289576 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289799 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289820 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.289832 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c8ed3bb8-646b-4db7-9b7f-556f5625d104-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.290117 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c8ed3bb8-646b-4db7-9b7f-556f5625d104" (UID: "c8ed3bb8-646b-4db7-9b7f-556f5625d104"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.290314 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-scripts" (OuterVolumeSpecName: "scripts") pod "c8ed3bb8-646b-4db7-9b7f-556f5625d104" (UID: "c8ed3bb8-646b-4db7-9b7f-556f5625d104"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.299850 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ed3bb8-646b-4db7-9b7f-556f5625d104-kube-api-access-vgz9v" (OuterVolumeSpecName: "kube-api-access-vgz9v") pod "c8ed3bb8-646b-4db7-9b7f-556f5625d104" (UID: "c8ed3bb8-646b-4db7-9b7f-556f5625d104"). InnerVolumeSpecName "kube-api-access-vgz9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.391354 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.391405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.391463 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.391556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.391616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9t8j\" (UniqueName: \"kubernetes.io/projected/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-kube-api-access-f9t8j\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.391695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-config\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.391880 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.392352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.392510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.392841 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8ed3bb8-646b-4db7-9b7f-556f5625d104-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.392864 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgz9v\" (UniqueName: \"kubernetes.io/projected/c8ed3bb8-646b-4db7-9b7f-556f5625d104-kube-api-access-vgz9v\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.393091 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.393203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.393480 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-config\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.408457 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9t8j\" (UniqueName: \"kubernetes.io/projected/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-kube-api-access-f9t8j\") pod \"dnsmasq-dns-77585f5f8c-zmh58\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.500135 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.795957 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-q7bdj" Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.796304 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz-config-q7bdj" event={"ID":"c8ed3bb8-646b-4db7-9b7f-556f5625d104","Type":"ContainerDied","Data":"d6f7280bb0e86d631b4856387bc3296bafbaad3a6fd983b0a0f44724887cbe3c"} Dec 05 11:29:42 crc kubenswrapper[4809]: I1205 11:29:42.796337 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6f7280bb0e86d631b4856387bc3296bafbaad3a6fd983b0a0f44724887cbe3c" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.023660 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-zmh58"] Dec 05 11:29:43 crc kubenswrapper[4809]: W1205 11:29:43.024913 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb41fa1fd_0666_4f63_b6f5_ee82fcb0031a.slice/crio-b156f5e0bb2b10bcf3ce3b1d0f2ff368bcd2dcf3c4f49fd0c90903217a202582 WatchSource:0}: Error finding container b156f5e0bb2b10bcf3ce3b1d0f2ff368bcd2dcf3c4f49fd0c90903217a202582: Status 404 returned error can't find the container with id b156f5e0bb2b10bcf3ce3b1d0f2ff368bcd2dcf3c4f49fd0c90903217a202582 Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.323826 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fg4wz-config-q7bdj"] Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.337020 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fg4wz-config-q7bdj"] Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.384926 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fg4wz-config-7hvm4"] Dec 05 11:29:43 crc kubenswrapper[4809]: E1205 11:29:43.385354 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ed3bb8-646b-4db7-9b7f-556f5625d104" containerName="ovn-config" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.385379 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ed3bb8-646b-4db7-9b7f-556f5625d104" containerName="ovn-config" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.388975 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ed3bb8-646b-4db7-9b7f-556f5625d104" containerName="ovn-config" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.389724 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.393846 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.486853 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fg4wz-config-7hvm4"] Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.513977 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b692\" (UniqueName: \"kubernetes.io/projected/9eb02312-519c-4099-bafe-8b47d65a1f94-kube-api-access-5b692\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.514065 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.514090 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-additional-scripts\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.514211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-log-ovn\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.514411 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run-ovn\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.514447 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-scripts\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run-ovn\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616370 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-scripts\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616408 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b692\" (UniqueName: \"kubernetes.io/projected/9eb02312-519c-4099-bafe-8b47d65a1f94-kube-api-access-5b692\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616484 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-additional-scripts\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-log-ovn\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616663 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run-ovn\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616672 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-log-ovn\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.616698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.617266 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-additional-scripts\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.618334 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-scripts\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.634719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b692\" (UniqueName: \"kubernetes.io/projected/9eb02312-519c-4099-bafe-8b47d65a1f94-kube-api-access-5b692\") pod \"ovn-controller-fg4wz-config-7hvm4\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.713176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.823032 4809 generic.go:334] "Generic (PLEG): container finished" podID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerID="ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7" exitCode=0 Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.823079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" event={"ID":"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a","Type":"ContainerDied","Data":"ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7"} Dec 05 11:29:43 crc kubenswrapper[4809]: I1205 11:29:43.823120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" event={"ID":"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a","Type":"ContainerStarted","Data":"b156f5e0bb2b10bcf3ce3b1d0f2ff368bcd2dcf3c4f49fd0c90903217a202582"} Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.163572 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fg4wz-config-7hvm4"] Dec 05 11:29:44 crc kubenswrapper[4809]: W1205 11:29:44.170772 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eb02312_519c_4099_bafe_8b47d65a1f94.slice/crio-9b774cc3b5059fb14a104bd21238856fe518b95687022ebc0835dca7485dbd46 WatchSource:0}: Error finding container 9b774cc3b5059fb14a104bd21238856fe518b95687022ebc0835dca7485dbd46: Status 404 returned error can't find the container with id 9b774cc3b5059fb14a104bd21238856fe518b95687022ebc0835dca7485dbd46 Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.837554 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" event={"ID":"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a","Type":"ContainerStarted","Data":"8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc"} Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.838051 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.840308 4809 generic.go:334] "Generic (PLEG): container finished" podID="9eb02312-519c-4099-bafe-8b47d65a1f94" containerID="aad7adfa2eb8eca2b71c201df7cb8a7e5b0a169a22eecd2c354e23760da7454b" exitCode=0 Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.840336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz-config-7hvm4" event={"ID":"9eb02312-519c-4099-bafe-8b47d65a1f94","Type":"ContainerDied","Data":"aad7adfa2eb8eca2b71c201df7cb8a7e5b0a169a22eecd2c354e23760da7454b"} Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.840351 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz-config-7hvm4" event={"ID":"9eb02312-519c-4099-bafe-8b47d65a1f94","Type":"ContainerStarted","Data":"9b774cc3b5059fb14a104bd21238856fe518b95687022ebc0835dca7485dbd46"} Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.860215 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" podStartSLOduration=2.860194707 podStartE2EDuration="2.860194707s" podCreationTimestamp="2025-12-05 11:29:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:44.856817589 +0000 UTC m=+1280.247794167" watchObservedRunningTime="2025-12-05 11:29:44.860194707 +0000 UTC m=+1280.251171265" Dec 05 11:29:44 crc kubenswrapper[4809]: I1205 11:29:44.893783 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ed3bb8-646b-4db7-9b7f-556f5625d104" path="/var/lib/kubelet/pods/c8ed3bb8-646b-4db7-9b7f-556f5625d104/volumes" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.215542 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.377437 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-log-ovn\") pod \"9eb02312-519c-4099-bafe-8b47d65a1f94\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.377577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-scripts\") pod \"9eb02312-519c-4099-bafe-8b47d65a1f94\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.377687 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-additional-scripts\") pod \"9eb02312-519c-4099-bafe-8b47d65a1f94\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.377686 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9eb02312-519c-4099-bafe-8b47d65a1f94" (UID: "9eb02312-519c-4099-bafe-8b47d65a1f94"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.377763 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b692\" (UniqueName: \"kubernetes.io/projected/9eb02312-519c-4099-bafe-8b47d65a1f94-kube-api-access-5b692\") pod \"9eb02312-519c-4099-bafe-8b47d65a1f94\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.377792 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run\") pod \"9eb02312-519c-4099-bafe-8b47d65a1f94\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.377870 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run-ovn\") pod \"9eb02312-519c-4099-bafe-8b47d65a1f94\" (UID: \"9eb02312-519c-4099-bafe-8b47d65a1f94\") " Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.378027 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run" (OuterVolumeSpecName: "var-run") pod "9eb02312-519c-4099-bafe-8b47d65a1f94" (UID: "9eb02312-519c-4099-bafe-8b47d65a1f94"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.378157 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9eb02312-519c-4099-bafe-8b47d65a1f94" (UID: "9eb02312-519c-4099-bafe-8b47d65a1f94"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.378720 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.378761 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.378778 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9eb02312-519c-4099-bafe-8b47d65a1f94-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.379940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9eb02312-519c-4099-bafe-8b47d65a1f94" (UID: "9eb02312-519c-4099-bafe-8b47d65a1f94"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.380410 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-scripts" (OuterVolumeSpecName: "scripts") pod "9eb02312-519c-4099-bafe-8b47d65a1f94" (UID: "9eb02312-519c-4099-bafe-8b47d65a1f94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.384350 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb02312-519c-4099-bafe-8b47d65a1f94-kube-api-access-5b692" (OuterVolumeSpecName: "kube-api-access-5b692") pod "9eb02312-519c-4099-bafe-8b47d65a1f94" (UID: "9eb02312-519c-4099-bafe-8b47d65a1f94"). InnerVolumeSpecName "kube-api-access-5b692". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.480654 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.480711 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb02312-519c-4099-bafe-8b47d65a1f94-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.480735 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b692\" (UniqueName: \"kubernetes.io/projected/9eb02312-519c-4099-bafe-8b47d65a1f94-kube-api-access-5b692\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.876813 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz-config-7hvm4" Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.885083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz-config-7hvm4" event={"ID":"9eb02312-519c-4099-bafe-8b47d65a1f94","Type":"ContainerDied","Data":"9b774cc3b5059fb14a104bd21238856fe518b95687022ebc0835dca7485dbd46"} Dec 05 11:29:46 crc kubenswrapper[4809]: I1205 11:29:46.885138 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b774cc3b5059fb14a104bd21238856fe518b95687022ebc0835dca7485dbd46" Dec 05 11:29:47 crc kubenswrapper[4809]: I1205 11:29:47.344984 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fg4wz-config-7hvm4"] Dec 05 11:29:47 crc kubenswrapper[4809]: I1205 11:29:47.353347 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fg4wz-config-7hvm4"] Dec 05 11:29:48 crc kubenswrapper[4809]: I1205 11:29:48.886701 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb02312-519c-4099-bafe-8b47d65a1f94" path="/var/lib/kubelet/pods/9eb02312-519c-4099-bafe-8b47d65a1f94/volumes" Dec 05 11:29:48 crc kubenswrapper[4809]: I1205 11:29:48.895063 4809 generic.go:334] "Generic (PLEG): container finished" podID="b4e08826-2c05-4f4f-b899-7a87f44e51d5" containerID="32e6b77cca37dcdbc635d76201fe078ccada3173763f886d6471946dd5311728" exitCode=0 Dec 05 11:29:48 crc kubenswrapper[4809]: I1205 11:29:48.895405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5mwp9" event={"ID":"b4e08826-2c05-4f4f-b899-7a87f44e51d5","Type":"ContainerDied","Data":"32e6b77cca37dcdbc635d76201fe078ccada3173763f886d6471946dd5311728"} Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.379797 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.477863 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-combined-ca-bundle\") pod \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.477946 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw4bm\" (UniqueName: \"kubernetes.io/projected/b4e08826-2c05-4f4f-b899-7a87f44e51d5-kube-api-access-nw4bm\") pod \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.478034 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-db-sync-config-data\") pod \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.478233 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-config-data\") pod \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\" (UID: \"b4e08826-2c05-4f4f-b899-7a87f44e51d5\") " Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.483547 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e08826-2c05-4f4f-b899-7a87f44e51d5-kube-api-access-nw4bm" (OuterVolumeSpecName: "kube-api-access-nw4bm") pod "b4e08826-2c05-4f4f-b899-7a87f44e51d5" (UID: "b4e08826-2c05-4f4f-b899-7a87f44e51d5"). InnerVolumeSpecName "kube-api-access-nw4bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.487089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b4e08826-2c05-4f4f-b899-7a87f44e51d5" (UID: "b4e08826-2c05-4f4f-b899-7a87f44e51d5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.519331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4e08826-2c05-4f4f-b899-7a87f44e51d5" (UID: "b4e08826-2c05-4f4f-b899-7a87f44e51d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.548073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-config-data" (OuterVolumeSpecName: "config-data") pod "b4e08826-2c05-4f4f-b899-7a87f44e51d5" (UID: "b4e08826-2c05-4f4f-b899-7a87f44e51d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.580009 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.580049 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.580084 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw4bm\" (UniqueName: \"kubernetes.io/projected/b4e08826-2c05-4f4f-b899-7a87f44e51d5-kube-api-access-nw4bm\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.580095 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e08826-2c05-4f4f-b899-7a87f44e51d5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.918496 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5mwp9" event={"ID":"b4e08826-2c05-4f4f-b899-7a87f44e51d5","Type":"ContainerDied","Data":"485da7f3ff051ba72459adb79eb97be8276176e361305c2af90c641725b0ff71"} Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.918543 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="485da7f3ff051ba72459adb79eb97be8276176e361305c2af90c641725b0ff71" Dec 05 11:29:50 crc kubenswrapper[4809]: I1205 11:29:50.918586 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5mwp9" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.326496 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-zmh58"] Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.327110 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" podUID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerName="dnsmasq-dns" containerID="cri-o://8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc" gracePeriod=10 Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.328843 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.350379 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jmvtl"] Dec 05 11:29:51 crc kubenswrapper[4809]: E1205 11:29:51.350730 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb02312-519c-4099-bafe-8b47d65a1f94" containerName="ovn-config" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.350745 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb02312-519c-4099-bafe-8b47d65a1f94" containerName="ovn-config" Dec 05 11:29:51 crc kubenswrapper[4809]: E1205 11:29:51.350766 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e08826-2c05-4f4f-b899-7a87f44e51d5" containerName="glance-db-sync" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.350773 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e08826-2c05-4f4f-b899-7a87f44e51d5" containerName="glance-db-sync" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.350917 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb02312-519c-4099-bafe-8b47d65a1f94" containerName="ovn-config" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.350940 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e08826-2c05-4f4f-b899-7a87f44e51d5" containerName="glance-db-sync" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.351771 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.378717 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jmvtl"] Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.495376 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-config\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.495433 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.495503 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7wlr\" (UniqueName: \"kubernetes.io/projected/b8721dce-921b-4ac4-ae84-94815ae99e67-kube-api-access-f7wlr\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.495527 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.495587 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.495617 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.597603 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.597735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7wlr\" (UniqueName: \"kubernetes.io/projected/b8721dce-921b-4ac4-ae84-94815ae99e67-kube-api-access-f7wlr\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.597760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.597814 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.597843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.597876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-config\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.598808 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.598852 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-config\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.598852 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.598902 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.599141 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.619605 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7wlr\" (UniqueName: \"kubernetes.io/projected/b8721dce-921b-4ac4-ae84-94815ae99e67-kube-api-access-f7wlr\") pod \"dnsmasq-dns-7ff5475cc9-jmvtl\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.672408 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:51 crc kubenswrapper[4809]: I1205 11:29:51.852862 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.143173 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jmvtl"] Dec 05 11:29:52 crc kubenswrapper[4809]: W1205 11:29:52.158195 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8721dce_921b_4ac4_ae84_94815ae99e67.slice/crio-31cbd4b24d9a8faa6963097c0cae290acbb66f065ef52da499c7bd9fd7337906 WatchSource:0}: Error finding container 31cbd4b24d9a8faa6963097c0cae290acbb66f065ef52da499c7bd9fd7337906: Status 404 returned error can't find the container with id 31cbd4b24d9a8faa6963097c0cae290acbb66f065ef52da499c7bd9fd7337906 Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.162853 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-86p6h"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.163903 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.193612 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-86p6h"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.273775 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-18b7-account-create-update-99wvm"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.274942 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.282903 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.285813 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-bxdwk"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.286910 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.296764 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bxdwk"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.308610 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9868714a-454b-4647-9b49-58b91e834fb3-operator-scripts\") pod \"cinder-db-create-86p6h\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.308695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m566d\" (UniqueName: \"kubernetes.io/projected/9868714a-454b-4647-9b49-58b91e834fb3-kube-api-access-m566d\") pod \"cinder-db-create-86p6h\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.308793 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-18b7-account-create-update-99wvm"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.407807 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.410287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m566d\" (UniqueName: \"kubernetes.io/projected/9868714a-454b-4647-9b49-58b91e834fb3-kube-api-access-m566d\") pod \"cinder-db-create-86p6h\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.410361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d0bc12-98af-4fc5-bad7-509b69c92f79-operator-scripts\") pod \"barbican-db-create-bxdwk\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.410434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqqmn\" (UniqueName: \"kubernetes.io/projected/3a3f0d8a-08af-4640-b937-123d6e3405d7-kube-api-access-jqqmn\") pod \"cinder-18b7-account-create-update-99wvm\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.410479 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a3f0d8a-08af-4640-b937-123d6e3405d7-operator-scripts\") pod \"cinder-18b7-account-create-update-99wvm\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.410520 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9868714a-454b-4647-9b49-58b91e834fb3-operator-scripts\") pod \"cinder-db-create-86p6h\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.410556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnvkz\" (UniqueName: \"kubernetes.io/projected/30d0bc12-98af-4fc5-bad7-509b69c92f79-kube-api-access-hnvkz\") pod \"barbican-db-create-bxdwk\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.411755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9868714a-454b-4647-9b49-58b91e834fb3-operator-scripts\") pod \"cinder-db-create-86p6h\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.422169 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-92b5-account-create-update-8snn6"] Dec 05 11:29:52 crc kubenswrapper[4809]: E1205 11:29:52.422958 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerName="dnsmasq-dns" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.422985 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerName="dnsmasq-dns" Dec 05 11:29:52 crc kubenswrapper[4809]: E1205 11:29:52.423008 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerName="init" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.423016 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerName="init" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.423328 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerName="dnsmasq-dns" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.424329 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.428779 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.428965 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-92b5-account-create-update-8snn6"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.434888 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m566d\" (UniqueName: \"kubernetes.io/projected/9868714a-454b-4647-9b49-58b91e834fb3-kube-api-access-m566d\") pod \"cinder-db-create-86p6h\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.511130 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-swift-storage-0\") pod \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.511680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9t8j\" (UniqueName: \"kubernetes.io/projected/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-kube-api-access-f9t8j\") pod \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.511722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-svc\") pod \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.512817 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-sb\") pod \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.512902 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-nb\") pod \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.512966 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-config\") pod \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\" (UID: \"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a\") " Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.513289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqqmn\" (UniqueName: \"kubernetes.io/projected/3a3f0d8a-08af-4640-b937-123d6e3405d7-kube-api-access-jqqmn\") pod \"cinder-18b7-account-create-update-99wvm\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.513339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a3f0d8a-08af-4640-b937-123d6e3405d7-operator-scripts\") pod \"cinder-18b7-account-create-update-99wvm\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.513613 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnvkz\" (UniqueName: \"kubernetes.io/projected/30d0bc12-98af-4fc5-bad7-509b69c92f79-kube-api-access-hnvkz\") pod \"barbican-db-create-bxdwk\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.513780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d0bc12-98af-4fc5-bad7-509b69c92f79-operator-scripts\") pod \"barbican-db-create-bxdwk\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.515347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d0bc12-98af-4fc5-bad7-509b69c92f79-operator-scripts\") pod \"barbican-db-create-bxdwk\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.516278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a3f0d8a-08af-4640-b937-123d6e3405d7-operator-scripts\") pod \"cinder-18b7-account-create-update-99wvm\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.518185 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-kube-api-access-f9t8j" (OuterVolumeSpecName: "kube-api-access-f9t8j") pod "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" (UID: "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a"). InnerVolumeSpecName "kube-api-access-f9t8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.513844 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tqlb\" (UniqueName: \"kubernetes.io/projected/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-kube-api-access-6tqlb\") pod \"barbican-92b5-account-create-update-8snn6\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.522804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-operator-scripts\") pod \"barbican-92b5-account-create-update-8snn6\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.522976 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9t8j\" (UniqueName: \"kubernetes.io/projected/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-kube-api-access-f9t8j\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.540694 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-22pc6"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.543461 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.551100 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.552274 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dnjt6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.552760 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.553237 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.553354 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.571135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqqmn\" (UniqueName: \"kubernetes.io/projected/3a3f0d8a-08af-4640-b937-123d6e3405d7-kube-api-access-jqqmn\") pod \"cinder-18b7-account-create-update-99wvm\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.571213 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-22pc6"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.572768 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnvkz\" (UniqueName: \"kubernetes.io/projected/30d0bc12-98af-4fc5-bad7-509b69c92f79-kube-api-access-hnvkz\") pod \"barbican-db-create-bxdwk\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.579975 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-2nbmc"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.580912 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.593955 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2nbmc"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.607577 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" (UID: "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.610847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.617855 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" (UID: "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.639017 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.639701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-operator-scripts\") pod \"barbican-92b5-account-create-update-8snn6\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.639830 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-combined-ca-bundle\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.639895 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-config-data\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.639932 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t42w\" (UniqueName: \"kubernetes.io/projected/5b2d9fda-bef0-4a73-84aa-c793181811d9-kube-api-access-8t42w\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.640001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tqlb\" (UniqueName: \"kubernetes.io/projected/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-kube-api-access-6tqlb\") pod \"barbican-92b5-account-create-update-8snn6\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.640114 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.640126 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.640974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-operator-scripts\") pod \"barbican-92b5-account-create-update-8snn6\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.642989 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" (UID: "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.672547 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" (UID: "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.679354 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-config" (OuterVolumeSpecName: "config") pod "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" (UID: "b41fa1fd-0666-4f63-b6f5-ee82fcb0031a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.681692 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tqlb\" (UniqueName: \"kubernetes.io/projected/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-kube-api-access-6tqlb\") pod \"barbican-92b5-account-create-update-8snn6\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.698075 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ee61-account-create-update-d8glr"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.699142 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.703223 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.717269 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ee61-account-create-update-d8glr"] Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.724715 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742426 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-combined-ca-bundle\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742505 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-config-data\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t42w\" (UniqueName: \"kubernetes.io/projected/5b2d9fda-bef0-4a73-84aa-c793181811d9-kube-api-access-8t42w\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742618 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbs66\" (UniqueName: \"kubernetes.io/projected/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-kube-api-access-qbs66\") pod \"neutron-db-create-2nbmc\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742675 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-operator-scripts\") pod \"neutron-db-create-2nbmc\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742803 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742819 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.742830 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.750483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-config-data\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.769298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-combined-ca-bundle\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.774537 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t42w\" (UniqueName: \"kubernetes.io/projected/5b2d9fda-bef0-4a73-84aa-c793181811d9-kube-api-access-8t42w\") pod \"keystone-db-sync-22pc6\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.798003 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.843901 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-operator-scripts\") pod \"neutron-db-create-2nbmc\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.844035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tjlm\" (UniqueName: \"kubernetes.io/projected/ad966040-b1a2-448d-b673-aeada07d064f-kube-api-access-7tjlm\") pod \"neutron-ee61-account-create-update-d8glr\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.844126 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad966040-b1a2-448d-b673-aeada07d064f-operator-scripts\") pod \"neutron-ee61-account-create-update-d8glr\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.844185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbs66\" (UniqueName: \"kubernetes.io/projected/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-kube-api-access-qbs66\") pod \"neutron-db-create-2nbmc\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.844886 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-operator-scripts\") pod \"neutron-db-create-2nbmc\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.868335 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbs66\" (UniqueName: \"kubernetes.io/projected/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-kube-api-access-qbs66\") pod \"neutron-db-create-2nbmc\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.941361 4809 generic.go:334] "Generic (PLEG): container finished" podID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerID="f6b8afa57cce5400518588e578205bb1cb1cdecff67726b8da835d625cf70a77" exitCode=0 Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.941433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" event={"ID":"b8721dce-921b-4ac4-ae84-94815ae99e67","Type":"ContainerDied","Data":"f6b8afa57cce5400518588e578205bb1cb1cdecff67726b8da835d625cf70a77"} Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.941462 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" event={"ID":"b8721dce-921b-4ac4-ae84-94815ae99e67","Type":"ContainerStarted","Data":"31cbd4b24d9a8faa6963097c0cae290acbb66f065ef52da499c7bd9fd7337906"} Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.945460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad966040-b1a2-448d-b673-aeada07d064f-operator-scripts\") pod \"neutron-ee61-account-create-update-d8glr\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.945578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tjlm\" (UniqueName: \"kubernetes.io/projected/ad966040-b1a2-448d-b673-aeada07d064f-kube-api-access-7tjlm\") pod \"neutron-ee61-account-create-update-d8glr\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.946436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad966040-b1a2-448d-b673-aeada07d064f-operator-scripts\") pod \"neutron-ee61-account-create-update-d8glr\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.949408 4809 generic.go:334] "Generic (PLEG): container finished" podID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" containerID="8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc" exitCode=0 Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.949437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" event={"ID":"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a","Type":"ContainerDied","Data":"8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc"} Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.949460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" event={"ID":"b41fa1fd-0666-4f63-b6f5-ee82fcb0031a","Type":"ContainerDied","Data":"b156f5e0bb2b10bcf3ce3b1d0f2ff368bcd2dcf3c4f49fd0c90903217a202582"} Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.949478 4809 scope.go:117] "RemoveContainer" containerID="8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.949590 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-zmh58" Dec 05 11:29:52 crc kubenswrapper[4809]: I1205 11:29:52.995887 4809 scope.go:117] "RemoveContainer" containerID="ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.007663 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-zmh58"] Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.009493 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tjlm\" (UniqueName: \"kubernetes.io/projected/ad966040-b1a2-448d-b673-aeada07d064f-kube-api-access-7tjlm\") pod \"neutron-ee61-account-create-update-d8glr\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.014299 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-zmh58"] Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.022688 4809 scope.go:117] "RemoveContainer" containerID="8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc" Dec 05 11:29:53 crc kubenswrapper[4809]: E1205 11:29:53.023620 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc\": container with ID starting with 8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc not found: ID does not exist" containerID="8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.023687 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc"} err="failed to get container status \"8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc\": rpc error: code = NotFound desc = could not find container \"8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc\": container with ID starting with 8a59d4daa26f67af109f7c032b00b82877f7bf82d8afe1df0582e8e61278cffc not found: ID does not exist" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.023708 4809 scope.go:117] "RemoveContainer" containerID="ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7" Dec 05 11:29:53 crc kubenswrapper[4809]: E1205 11:29:53.024018 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7\": container with ID starting with ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7 not found: ID does not exist" containerID="ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.024060 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7"} err="failed to get container status \"ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7\": rpc error: code = NotFound desc = could not find container \"ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7\": container with ID starting with ea3ca49452ea6b9098d6b54ebacc047b2bc585d7d0e461a0bc36012093953ae7 not found: ID does not exist" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.038719 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22pc6" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.042317 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.057874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.075581 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bxdwk"] Dec 05 11:29:53 crc kubenswrapper[4809]: W1205 11:29:53.089241 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30d0bc12_98af_4fc5_bad7_509b69c92f79.slice/crio-cc9a743639806c868c6771672c07a511348985184662f05e772fde6ed849924e WatchSource:0}: Error finding container cc9a743639806c868c6771672c07a511348985184662f05e772fde6ed849924e: Status 404 returned error can't find the container with id cc9a743639806c868c6771672c07a511348985184662f05e772fde6ed849924e Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.153919 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-18b7-account-create-update-99wvm"] Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.202938 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-86p6h"] Dec 05 11:29:53 crc kubenswrapper[4809]: I1205 11:29:53.652912 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-92b5-account-create-update-8snn6"] Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:53.971564 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-22pc6"] Dec 05 11:29:54 crc kubenswrapper[4809]: W1205 11:29:53.972979 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b2d9fda_bef0_4a73_84aa_c793181811d9.slice/crio-e5212bd4f5c7fa3d8ac934217fc6c4f33206fa64473d0f18adb7e9037197d43a WatchSource:0}: Error finding container e5212bd4f5c7fa3d8ac934217fc6c4f33206fa64473d0f18adb7e9037197d43a: Status 404 returned error can't find the container with id e5212bd4f5c7fa3d8ac934217fc6c4f33206fa64473d0f18adb7e9037197d43a Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:53.975565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" event={"ID":"b8721dce-921b-4ac4-ae84-94815ae99e67","Type":"ContainerStarted","Data":"f615d4be35da2629c2d18907ba7e6166ab3ef4c48f963d7e1f9a7707bd8f6399"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:53.976964 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:53.981601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bxdwk" event={"ID":"30d0bc12-98af-4fc5-bad7-509b69c92f79","Type":"ContainerStarted","Data":"d0a4c11aa9fc03b76e0161912fbe85f1f20894ffb26f11d0e3172a02fc7f3658"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:53.981653 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bxdwk" event={"ID":"30d0bc12-98af-4fc5-bad7-509b69c92f79","Type":"ContainerStarted","Data":"cc9a743639806c868c6771672c07a511348985184662f05e772fde6ed849924e"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:53.989168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-86p6h" event={"ID":"9868714a-454b-4647-9b49-58b91e834fb3","Type":"ContainerStarted","Data":"0be7f476d8685657cf0bb6c394823faf079fd3c3df299ee0d38d4ac3075cf7ae"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:53.989213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-86p6h" event={"ID":"9868714a-454b-4647-9b49-58b91e834fb3","Type":"ContainerStarted","Data":"2bb42e41ef5f94b32ee1ffd2f41ee315fcf92f83a8c3ea4390c95ee22c40f7b0"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.015881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-92b5-account-create-update-8snn6" event={"ID":"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2","Type":"ContainerStarted","Data":"b2f329660098f5d914595479a66fbd5e0b73b1633354e05daff2ff0d359489c9"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.015931 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-92b5-account-create-update-8snn6" event={"ID":"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2","Type":"ContainerStarted","Data":"20b45ec4f1af5ab6eeeb89bffa9346252cad436165a5d85a07c21d799f672754"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.019349 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-18b7-account-create-update-99wvm" event={"ID":"3a3f0d8a-08af-4640-b937-123d6e3405d7","Type":"ContainerStarted","Data":"f74b9b84031262b69b17ddf856c579346742b85b32806bca97cf70ab2d4209da"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.019383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-18b7-account-create-update-99wvm" event={"ID":"3a3f0d8a-08af-4640-b937-123d6e3405d7","Type":"ContainerStarted","Data":"d05d7558342f4c4d2305a568b6b83fb6ec2ce4b9d0c163ec7182e3671f10ce1d"} Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.032966 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" podStartSLOduration=3.032936807 podStartE2EDuration="3.032936807s" podCreationTimestamp="2025-12-05 11:29:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:54.022666818 +0000 UTC m=+1289.413643406" watchObservedRunningTime="2025-12-05 11:29:54.032936807 +0000 UTC m=+1289.423913375" Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.047473 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-92b5-account-create-update-8snn6" podStartSLOduration=2.047449857 podStartE2EDuration="2.047449857s" podCreationTimestamp="2025-12-05 11:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:54.042519998 +0000 UTC m=+1289.433496566" watchObservedRunningTime="2025-12-05 11:29:54.047449857 +0000 UTC m=+1289.438426415" Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.068847 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-bxdwk" podStartSLOduration=2.068820826 podStartE2EDuration="2.068820826s" podCreationTimestamp="2025-12-05 11:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:54.061070073 +0000 UTC m=+1289.452046651" watchObservedRunningTime="2025-12-05 11:29:54.068820826 +0000 UTC m=+1289.459797384" Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.088572 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-86p6h" podStartSLOduration=2.0885457020000002 podStartE2EDuration="2.088545702s" podCreationTimestamp="2025-12-05 11:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:54.078769066 +0000 UTC m=+1289.469745634" watchObservedRunningTime="2025-12-05 11:29:54.088545702 +0000 UTC m=+1289.479522270" Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.104685 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-18b7-account-create-update-99wvm" podStartSLOduration=2.104662704 podStartE2EDuration="2.104662704s" podCreationTimestamp="2025-12-05 11:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:29:54.102097256 +0000 UTC m=+1289.493073824" watchObservedRunningTime="2025-12-05 11:29:54.104662704 +0000 UTC m=+1289.495639272" Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.884722 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b41fa1fd-0666-4f63-b6f5-ee82fcb0031a" path="/var/lib/kubelet/pods/b41fa1fd-0666-4f63-b6f5-ee82fcb0031a/volumes" Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.885897 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ee61-account-create-update-d8glr"] Dec 05 11:29:54 crc kubenswrapper[4809]: I1205 11:29:54.895768 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2nbmc"] Dec 05 11:29:54 crc kubenswrapper[4809]: W1205 11:29:54.896301 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad966040_b1a2_448d_b673_aeada07d064f.slice/crio-575a5b12ad3a43d39d2c90cc04609140026201b6feb9b744d27d05c632647083 WatchSource:0}: Error finding container 575a5b12ad3a43d39d2c90cc04609140026201b6feb9b744d27d05c632647083: Status 404 returned error can't find the container with id 575a5b12ad3a43d39d2c90cc04609140026201b6feb9b744d27d05c632647083 Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.034263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22pc6" event={"ID":"5b2d9fda-bef0-4a73-84aa-c793181811d9","Type":"ContainerStarted","Data":"e5212bd4f5c7fa3d8ac934217fc6c4f33206fa64473d0f18adb7e9037197d43a"} Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.035840 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2nbmc" event={"ID":"fc29db8b-3c51-4b71-a984-41a1eb4a21c0","Type":"ContainerStarted","Data":"49a5697d9e8ccc059ab9c20c21c5691979eb14b1a4717a963421ade47985a83a"} Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.037195 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ee61-account-create-update-d8glr" event={"ID":"ad966040-b1a2-448d-b673-aeada07d064f","Type":"ContainerStarted","Data":"575a5b12ad3a43d39d2c90cc04609140026201b6feb9b744d27d05c632647083"} Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.040005 4809 generic.go:334] "Generic (PLEG): container finished" podID="8d1389c7-2aa9-4d9b-8070-a5e308a48ca2" containerID="b2f329660098f5d914595479a66fbd5e0b73b1633354e05daff2ff0d359489c9" exitCode=0 Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.040059 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-92b5-account-create-update-8snn6" event={"ID":"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2","Type":"ContainerDied","Data":"b2f329660098f5d914595479a66fbd5e0b73b1633354e05daff2ff0d359489c9"} Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.043299 4809 generic.go:334] "Generic (PLEG): container finished" podID="3a3f0d8a-08af-4640-b937-123d6e3405d7" containerID="f74b9b84031262b69b17ddf856c579346742b85b32806bca97cf70ab2d4209da" exitCode=0 Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.043349 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-18b7-account-create-update-99wvm" event={"ID":"3a3f0d8a-08af-4640-b937-123d6e3405d7","Type":"ContainerDied","Data":"f74b9b84031262b69b17ddf856c579346742b85b32806bca97cf70ab2d4209da"} Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.046201 4809 generic.go:334] "Generic (PLEG): container finished" podID="30d0bc12-98af-4fc5-bad7-509b69c92f79" containerID="d0a4c11aa9fc03b76e0161912fbe85f1f20894ffb26f11d0e3172a02fc7f3658" exitCode=0 Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.046255 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bxdwk" event={"ID":"30d0bc12-98af-4fc5-bad7-509b69c92f79","Type":"ContainerDied","Data":"d0a4c11aa9fc03b76e0161912fbe85f1f20894ffb26f11d0e3172a02fc7f3658"} Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.049089 4809 generic.go:334] "Generic (PLEG): container finished" podID="9868714a-454b-4647-9b49-58b91e834fb3" containerID="0be7f476d8685657cf0bb6c394823faf079fd3c3df299ee0d38d4ac3075cf7ae" exitCode=0 Dec 05 11:29:55 crc kubenswrapper[4809]: I1205 11:29:55.049851 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-86p6h" event={"ID":"9868714a-454b-4647-9b49-58b91e834fb3","Type":"ContainerDied","Data":"0be7f476d8685657cf0bb6c394823faf079fd3c3df299ee0d38d4ac3075cf7ae"} Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.061576 4809 generic.go:334] "Generic (PLEG): container finished" podID="ad966040-b1a2-448d-b673-aeada07d064f" containerID="e4918c7f89901cb6bc80954b1fdf3a854bdcb6bfe697fc6fc6ad89234ed88f5d" exitCode=0 Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.061774 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ee61-account-create-update-d8glr" event={"ID":"ad966040-b1a2-448d-b673-aeada07d064f","Type":"ContainerDied","Data":"e4918c7f89901cb6bc80954b1fdf3a854bdcb6bfe697fc6fc6ad89234ed88f5d"} Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.064681 4809 generic.go:334] "Generic (PLEG): container finished" podID="fc29db8b-3c51-4b71-a984-41a1eb4a21c0" containerID="e0b33876d9e64c7fd99f042a76d398b7db13a4e6ce06d131130aa7bed7b10815" exitCode=0 Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.064842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2nbmc" event={"ID":"fc29db8b-3c51-4b71-a984-41a1eb4a21c0","Type":"ContainerDied","Data":"e0b33876d9e64c7fd99f042a76d398b7db13a4e6ce06d131130aa7bed7b10815"} Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.454610 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.607043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-operator-scripts\") pod \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.607083 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tqlb\" (UniqueName: \"kubernetes.io/projected/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-kube-api-access-6tqlb\") pod \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\" (UID: \"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2\") " Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.608352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d1389c7-2aa9-4d9b-8070-a5e308a48ca2" (UID: "8d1389c7-2aa9-4d9b-8070-a5e308a48ca2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.617035 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-kube-api-access-6tqlb" (OuterVolumeSpecName: "kube-api-access-6tqlb") pod "8d1389c7-2aa9-4d9b-8070-a5e308a48ca2" (UID: "8d1389c7-2aa9-4d9b-8070-a5e308a48ca2"). InnerVolumeSpecName "kube-api-access-6tqlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.709145 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:56 crc kubenswrapper[4809]: I1205 11:29:56.709187 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tqlb\" (UniqueName: \"kubernetes.io/projected/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2-kube-api-access-6tqlb\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:57 crc kubenswrapper[4809]: I1205 11:29:57.081977 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-92b5-account-create-update-8snn6" event={"ID":"8d1389c7-2aa9-4d9b-8070-a5e308a48ca2","Type":"ContainerDied","Data":"20b45ec4f1af5ab6eeeb89bffa9346252cad436165a5d85a07c21d799f672754"} Dec 05 11:29:57 crc kubenswrapper[4809]: I1205 11:29:57.082025 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-92b5-account-create-update-8snn6" Dec 05 11:29:57 crc kubenswrapper[4809]: I1205 11:29:57.082040 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20b45ec4f1af5ab6eeeb89bffa9346252cad436165a5d85a07c21d799f672754" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.050069 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.071121 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.071511 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.106771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bxdwk" event={"ID":"30d0bc12-98af-4fc5-bad7-509b69c92f79","Type":"ContainerDied","Data":"cc9a743639806c868c6771672c07a511348985184662f05e772fde6ed849924e"} Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.106864 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc9a743639806c868c6771672c07a511348985184662f05e772fde6ed849924e" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.112952 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-86p6h" event={"ID":"9868714a-454b-4647-9b49-58b91e834fb3","Type":"ContainerDied","Data":"2bb42e41ef5f94b32ee1ffd2f41ee315fcf92f83a8c3ea4390c95ee22c40f7b0"} Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.113003 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bb42e41ef5f94b32ee1ffd2f41ee315fcf92f83a8c3ea4390c95ee22c40f7b0" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.113082 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-86p6h" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.118184 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2nbmc" event={"ID":"fc29db8b-3c51-4b71-a984-41a1eb4a21c0","Type":"ContainerDied","Data":"49a5697d9e8ccc059ab9c20c21c5691979eb14b1a4717a963421ade47985a83a"} Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.118229 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49a5697d9e8ccc059ab9c20c21c5691979eb14b1a4717a963421ade47985a83a" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.118309 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2nbmc" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.122952 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ee61-account-create-update-d8glr" event={"ID":"ad966040-b1a2-448d-b673-aeada07d064f","Type":"ContainerDied","Data":"575a5b12ad3a43d39d2c90cc04609140026201b6feb9b744d27d05c632647083"} Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.122978 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="575a5b12ad3a43d39d2c90cc04609140026201b6feb9b744d27d05c632647083" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.125846 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-18b7-account-create-update-99wvm" event={"ID":"3a3f0d8a-08af-4640-b937-123d6e3405d7","Type":"ContainerDied","Data":"d05d7558342f4c4d2305a568b6b83fb6ec2ce4b9d0c163ec7182e3671f10ce1d"} Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.125875 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d05d7558342f4c4d2305a568b6b83fb6ec2ce4b9d0c163ec7182e3671f10ce1d" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.125912 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-18b7-account-create-update-99wvm" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.127969 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.137795 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bxdwk" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.150505 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbs66\" (UniqueName: \"kubernetes.io/projected/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-kube-api-access-qbs66\") pod \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.150573 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m566d\" (UniqueName: \"kubernetes.io/projected/9868714a-454b-4647-9b49-58b91e834fb3-kube-api-access-m566d\") pod \"9868714a-454b-4647-9b49-58b91e834fb3\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.150691 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-operator-scripts\") pod \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\" (UID: \"fc29db8b-3c51-4b71-a984-41a1eb4a21c0\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.150722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqqmn\" (UniqueName: \"kubernetes.io/projected/3a3f0d8a-08af-4640-b937-123d6e3405d7-kube-api-access-jqqmn\") pod \"3a3f0d8a-08af-4640-b937-123d6e3405d7\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.150747 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a3f0d8a-08af-4640-b937-123d6e3405d7-operator-scripts\") pod \"3a3f0d8a-08af-4640-b937-123d6e3405d7\" (UID: \"3a3f0d8a-08af-4640-b937-123d6e3405d7\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.150808 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9868714a-454b-4647-9b49-58b91e834fb3-operator-scripts\") pod \"9868714a-454b-4647-9b49-58b91e834fb3\" (UID: \"9868714a-454b-4647-9b49-58b91e834fb3\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.151554 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9868714a-454b-4647-9b49-58b91e834fb3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9868714a-454b-4647-9b49-58b91e834fb3" (UID: "9868714a-454b-4647-9b49-58b91e834fb3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.151772 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc29db8b-3c51-4b71-a984-41a1eb4a21c0" (UID: "fc29db8b-3c51-4b71-a984-41a1eb4a21c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.152443 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3f0d8a-08af-4640-b937-123d6e3405d7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a3f0d8a-08af-4640-b937-123d6e3405d7" (UID: "3a3f0d8a-08af-4640-b937-123d6e3405d7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.159877 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3f0d8a-08af-4640-b937-123d6e3405d7-kube-api-access-jqqmn" (OuterVolumeSpecName: "kube-api-access-jqqmn") pod "3a3f0d8a-08af-4640-b937-123d6e3405d7" (UID: "3a3f0d8a-08af-4640-b937-123d6e3405d7"). InnerVolumeSpecName "kube-api-access-jqqmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.172567 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9868714a-454b-4647-9b49-58b91e834fb3-kube-api-access-m566d" (OuterVolumeSpecName: "kube-api-access-m566d") pod "9868714a-454b-4647-9b49-58b91e834fb3" (UID: "9868714a-454b-4647-9b49-58b91e834fb3"). InnerVolumeSpecName "kube-api-access-m566d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.172910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-kube-api-access-qbs66" (OuterVolumeSpecName: "kube-api-access-qbs66") pod "fc29db8b-3c51-4b71-a984-41a1eb4a21c0" (UID: "fc29db8b-3c51-4b71-a984-41a1eb4a21c0"). InnerVolumeSpecName "kube-api-access-qbs66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.252566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d0bc12-98af-4fc5-bad7-509b69c92f79-operator-scripts\") pod \"30d0bc12-98af-4fc5-bad7-509b69c92f79\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.252664 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tjlm\" (UniqueName: \"kubernetes.io/projected/ad966040-b1a2-448d-b673-aeada07d064f-kube-api-access-7tjlm\") pod \"ad966040-b1a2-448d-b673-aeada07d064f\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.252694 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad966040-b1a2-448d-b673-aeada07d064f-operator-scripts\") pod \"ad966040-b1a2-448d-b673-aeada07d064f\" (UID: \"ad966040-b1a2-448d-b673-aeada07d064f\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.252860 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnvkz\" (UniqueName: \"kubernetes.io/projected/30d0bc12-98af-4fc5-bad7-509b69c92f79-kube-api-access-hnvkz\") pod \"30d0bc12-98af-4fc5-bad7-509b69c92f79\" (UID: \"30d0bc12-98af-4fc5-bad7-509b69c92f79\") " Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253321 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m566d\" (UniqueName: \"kubernetes.io/projected/9868714a-454b-4647-9b49-58b91e834fb3-kube-api-access-m566d\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253343 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253356 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqqmn\" (UniqueName: \"kubernetes.io/projected/3a3f0d8a-08af-4640-b937-123d6e3405d7-kube-api-access-jqqmn\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253368 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a3f0d8a-08af-4640-b937-123d6e3405d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad966040-b1a2-448d-b673-aeada07d064f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad966040-b1a2-448d-b673-aeada07d064f" (UID: "ad966040-b1a2-448d-b673-aeada07d064f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d0bc12-98af-4fc5-bad7-509b69c92f79-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30d0bc12-98af-4fc5-bad7-509b69c92f79" (UID: "30d0bc12-98af-4fc5-bad7-509b69c92f79"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253377 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9868714a-454b-4647-9b49-58b91e834fb3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.253434 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbs66\" (UniqueName: \"kubernetes.io/projected/fc29db8b-3c51-4b71-a984-41a1eb4a21c0-kube-api-access-qbs66\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.255652 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad966040-b1a2-448d-b673-aeada07d064f-kube-api-access-7tjlm" (OuterVolumeSpecName: "kube-api-access-7tjlm") pod "ad966040-b1a2-448d-b673-aeada07d064f" (UID: "ad966040-b1a2-448d-b673-aeada07d064f"). InnerVolumeSpecName "kube-api-access-7tjlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.256571 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d0bc12-98af-4fc5-bad7-509b69c92f79-kube-api-access-hnvkz" (OuterVolumeSpecName: "kube-api-access-hnvkz") pod "30d0bc12-98af-4fc5-bad7-509b69c92f79" (UID: "30d0bc12-98af-4fc5-bad7-509b69c92f79"). InnerVolumeSpecName "kube-api-access-hnvkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.355015 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d0bc12-98af-4fc5-bad7-509b69c92f79-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.355075 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tjlm\" (UniqueName: \"kubernetes.io/projected/ad966040-b1a2-448d-b673-aeada07d064f-kube-api-access-7tjlm\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.355092 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad966040-b1a2-448d-b673-aeada07d064f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:29:59 crc kubenswrapper[4809]: I1205 11:29:59.355103 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnvkz\" (UniqueName: \"kubernetes.io/projected/30d0bc12-98af-4fc5-bad7-509b69c92f79-kube-api-access-hnvkz\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.136350 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bxdwk" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.136389 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ee61-account-create-update-d8glr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.136318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22pc6" event={"ID":"5b2d9fda-bef0-4a73-84aa-c793181811d9","Type":"ContainerStarted","Data":"63cb23ada3d99ae3c4ef0234a641edf1c1ba2827cee6b00495704fb452a0053e"} Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.144482 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr"] Dec 05 11:30:00 crc kubenswrapper[4809]: E1205 11:30:00.148104 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad966040-b1a2-448d-b673-aeada07d064f" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148143 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad966040-b1a2-448d-b673-aeada07d064f" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: E1205 11:30:00.148164 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc29db8b-3c51-4b71-a984-41a1eb4a21c0" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148173 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc29db8b-3c51-4b71-a984-41a1eb4a21c0" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: E1205 11:30:00.148191 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9868714a-454b-4647-9b49-58b91e834fb3" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148197 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9868714a-454b-4647-9b49-58b91e834fb3" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: E1205 11:30:00.148209 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3f0d8a-08af-4640-b937-123d6e3405d7" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148216 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3f0d8a-08af-4640-b937-123d6e3405d7" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: E1205 11:30:00.148228 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1389c7-2aa9-4d9b-8070-a5e308a48ca2" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148237 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1389c7-2aa9-4d9b-8070-a5e308a48ca2" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: E1205 11:30:00.148262 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d0bc12-98af-4fc5-bad7-509b69c92f79" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148270 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d0bc12-98af-4fc5-bad7-509b69c92f79" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148455 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d0bc12-98af-4fc5-bad7-509b69c92f79" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148473 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad966040-b1a2-448d-b673-aeada07d064f" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148487 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9868714a-454b-4647-9b49-58b91e834fb3" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148495 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc29db8b-3c51-4b71-a984-41a1eb4a21c0" containerName="mariadb-database-create" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148505 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1389c7-2aa9-4d9b-8070-a5e308a48ca2" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.148518 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3f0d8a-08af-4640-b937-123d6e3405d7" containerName="mariadb-account-create-update" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.149187 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.153783 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.154868 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.163375 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr"] Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.190928 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-22pc6" podStartSLOduration=3.20444584 podStartE2EDuration="8.190909704s" podCreationTimestamp="2025-12-05 11:29:52 +0000 UTC" firstStartedPulling="2025-12-05 11:29:53.978901053 +0000 UTC m=+1289.369877621" lastFinishedPulling="2025-12-05 11:29:58.965364917 +0000 UTC m=+1294.356341485" observedRunningTime="2025-12-05 11:30:00.180642125 +0000 UTC m=+1295.571618683" watchObservedRunningTime="2025-12-05 11:30:00.190909704 +0000 UTC m=+1295.581886262" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.273803 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fjhd\" (UniqueName: \"kubernetes.io/projected/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-kube-api-access-2fjhd\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.273896 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-secret-volume\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.273995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-config-volume\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.375724 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-secret-volume\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.375823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-config-volume\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.375933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fjhd\" (UniqueName: \"kubernetes.io/projected/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-kube-api-access-2fjhd\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.377410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-config-volume\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.383252 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-secret-volume\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.395303 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fjhd\" (UniqueName: \"kubernetes.io/projected/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-kube-api-access-2fjhd\") pod \"collect-profiles-29415570-drknr\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.483811 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:00 crc kubenswrapper[4809]: I1205 11:30:00.981627 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr"] Dec 05 11:30:00 crc kubenswrapper[4809]: W1205 11:30:00.984202 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19ee7bf0_2a51_4926_b47d_c06c792c1fe9.slice/crio-871a6e0bdd331ca30a5f292c6d2a55f91b36fb46901cbc46fed7445b82c670a4 WatchSource:0}: Error finding container 871a6e0bdd331ca30a5f292c6d2a55f91b36fb46901cbc46fed7445b82c670a4: Status 404 returned error can't find the container with id 871a6e0bdd331ca30a5f292c6d2a55f91b36fb46901cbc46fed7445b82c670a4 Dec 05 11:30:01 crc kubenswrapper[4809]: I1205 11:30:01.147047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" event={"ID":"19ee7bf0-2a51-4926-b47d-c06c792c1fe9","Type":"ContainerStarted","Data":"871a6e0bdd331ca30a5f292c6d2a55f91b36fb46901cbc46fed7445b82c670a4"} Dec 05 11:30:01 crc kubenswrapper[4809]: I1205 11:30:01.673813 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:30:01 crc kubenswrapper[4809]: I1205 11:30:01.732770 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-68v9p"] Dec 05 11:30:01 crc kubenswrapper[4809]: I1205 11:30:01.733059 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-68v9p" podUID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerName="dnsmasq-dns" containerID="cri-o://e1ef014dab365de170de8bcc18bc67b0c2a056e8f731e71b7a9f0c26d47e7122" gracePeriod=10 Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.157765 4809 generic.go:334] "Generic (PLEG): container finished" podID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerID="e1ef014dab365de170de8bcc18bc67b0c2a056e8f731e71b7a9f0c26d47e7122" exitCode=0 Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.157880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-68v9p" event={"ID":"de639392-c7d7-4b7f-b894-0fb5d32f00a2","Type":"ContainerDied","Data":"e1ef014dab365de170de8bcc18bc67b0c2a056e8f731e71b7a9f0c26d47e7122"} Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.158214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-68v9p" event={"ID":"de639392-c7d7-4b7f-b894-0fb5d32f00a2","Type":"ContainerDied","Data":"e8595373053c2406cac8fbb2231f57767a9303d2feda538df1853c6a7c06cfe8"} Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.158231 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8595373053c2406cac8fbb2231f57767a9303d2feda538df1853c6a7c06cfe8" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.161161 4809 generic.go:334] "Generic (PLEG): container finished" podID="19ee7bf0-2a51-4926-b47d-c06c792c1fe9" containerID="37f4400738a589859e4d3607ebf099a53bd2eb5843646f4ec28456a932d2123d" exitCode=0 Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.161213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" event={"ID":"19ee7bf0-2a51-4926-b47d-c06c792c1fe9","Type":"ContainerDied","Data":"37f4400738a589859e4d3607ebf099a53bd2eb5843646f4ec28456a932d2123d"} Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.202567 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.310564 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-config\") pod \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.310686 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc\") pod \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.310721 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb\") pod \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.310824 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsxdg\" (UniqueName: \"kubernetes.io/projected/de639392-c7d7-4b7f-b894-0fb5d32f00a2-kube-api-access-xsxdg\") pod \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.310929 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-sb\") pod \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.351367 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de639392-c7d7-4b7f-b894-0fb5d32f00a2-kube-api-access-xsxdg" (OuterVolumeSpecName: "kube-api-access-xsxdg") pod "de639392-c7d7-4b7f-b894-0fb5d32f00a2" (UID: "de639392-c7d7-4b7f-b894-0fb5d32f00a2"). InnerVolumeSpecName "kube-api-access-xsxdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.415365 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de639392-c7d7-4b7f-b894-0fb5d32f00a2" (UID: "de639392-c7d7-4b7f-b894-0fb5d32f00a2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.422821 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de639392-c7d7-4b7f-b894-0fb5d32f00a2" (UID: "de639392-c7d7-4b7f-b894-0fb5d32f00a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.423711 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc\") pod \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.423937 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb\") pod \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\" (UID: \"de639392-c7d7-4b7f-b894-0fb5d32f00a2\") " Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.425597 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsxdg\" (UniqueName: \"kubernetes.io/projected/de639392-c7d7-4b7f-b894-0fb5d32f00a2-kube-api-access-xsxdg\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:02 crc kubenswrapper[4809]: W1205 11:30:02.423949 4809 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/de639392-c7d7-4b7f-b894-0fb5d32f00a2/volumes/kubernetes.io~configmap/dns-svc Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.426083 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de639392-c7d7-4b7f-b894-0fb5d32f00a2" (UID: "de639392-c7d7-4b7f-b894-0fb5d32f00a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:02 crc kubenswrapper[4809]: W1205 11:30:02.424007 4809 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/de639392-c7d7-4b7f-b894-0fb5d32f00a2/volumes/kubernetes.io~configmap/ovsdbserver-nb Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.426194 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de639392-c7d7-4b7f-b894-0fb5d32f00a2" (UID: "de639392-c7d7-4b7f-b894-0fb5d32f00a2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.438353 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de639392-c7d7-4b7f-b894-0fb5d32f00a2" (UID: "de639392-c7d7-4b7f-b894-0fb5d32f00a2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.438955 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-config" (OuterVolumeSpecName: "config") pod "de639392-c7d7-4b7f-b894-0fb5d32f00a2" (UID: "de639392-c7d7-4b7f-b894-0fb5d32f00a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.527735 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.527772 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.527782 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:02 crc kubenswrapper[4809]: I1205 11:30:02.527792 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de639392-c7d7-4b7f-b894-0fb5d32f00a2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.170234 4809 generic.go:334] "Generic (PLEG): container finished" podID="5b2d9fda-bef0-4a73-84aa-c793181811d9" containerID="63cb23ada3d99ae3c4ef0234a641edf1c1ba2827cee6b00495704fb452a0053e" exitCode=0 Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.170326 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22pc6" event={"ID":"5b2d9fda-bef0-4a73-84aa-c793181811d9","Type":"ContainerDied","Data":"63cb23ada3d99ae3c4ef0234a641edf1c1ba2827cee6b00495704fb452a0053e"} Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.170615 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-68v9p" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.223619 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-68v9p"] Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.223700 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-68v9p"] Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.506050 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.550102 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fjhd\" (UniqueName: \"kubernetes.io/projected/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-kube-api-access-2fjhd\") pod \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.550161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-secret-volume\") pod \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.550262 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-config-volume\") pod \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\" (UID: \"19ee7bf0-2a51-4926-b47d-c06c792c1fe9\") " Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.550999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-config-volume" (OuterVolumeSpecName: "config-volume") pod "19ee7bf0-2a51-4926-b47d-c06c792c1fe9" (UID: "19ee7bf0-2a51-4926-b47d-c06c792c1fe9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.555619 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-kube-api-access-2fjhd" (OuterVolumeSpecName: "kube-api-access-2fjhd") pod "19ee7bf0-2a51-4926-b47d-c06c792c1fe9" (UID: "19ee7bf0-2a51-4926-b47d-c06c792c1fe9"). InnerVolumeSpecName "kube-api-access-2fjhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.555905 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "19ee7bf0-2a51-4926-b47d-c06c792c1fe9" (UID: "19ee7bf0-2a51-4926-b47d-c06c792c1fe9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.651528 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.651563 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fjhd\" (UniqueName: \"kubernetes.io/projected/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-kube-api-access-2fjhd\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:03 crc kubenswrapper[4809]: I1205 11:30:03.651577 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19ee7bf0-2a51-4926-b47d-c06c792c1fe9-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.182195 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.182188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr" event={"ID":"19ee7bf0-2a51-4926-b47d-c06c792c1fe9","Type":"ContainerDied","Data":"871a6e0bdd331ca30a5f292c6d2a55f91b36fb46901cbc46fed7445b82c670a4"} Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.182233 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="871a6e0bdd331ca30a5f292c6d2a55f91b36fb46901cbc46fed7445b82c670a4" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.491393 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22pc6" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.565561 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-combined-ca-bundle\") pod \"5b2d9fda-bef0-4a73-84aa-c793181811d9\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.565889 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-config-data\") pod \"5b2d9fda-bef0-4a73-84aa-c793181811d9\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.565970 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t42w\" (UniqueName: \"kubernetes.io/projected/5b2d9fda-bef0-4a73-84aa-c793181811d9-kube-api-access-8t42w\") pod \"5b2d9fda-bef0-4a73-84aa-c793181811d9\" (UID: \"5b2d9fda-bef0-4a73-84aa-c793181811d9\") " Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.572285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b2d9fda-bef0-4a73-84aa-c793181811d9-kube-api-access-8t42w" (OuterVolumeSpecName: "kube-api-access-8t42w") pod "5b2d9fda-bef0-4a73-84aa-c793181811d9" (UID: "5b2d9fda-bef0-4a73-84aa-c793181811d9"). InnerVolumeSpecName "kube-api-access-8t42w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.596959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b2d9fda-bef0-4a73-84aa-c793181811d9" (UID: "5b2d9fda-bef0-4a73-84aa-c793181811d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.622143 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-config-data" (OuterVolumeSpecName: "config-data") pod "5b2d9fda-bef0-4a73-84aa-c793181811d9" (UID: "5b2d9fda-bef0-4a73-84aa-c793181811d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.667751 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.667786 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2d9fda-bef0-4a73-84aa-c793181811d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.667797 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t42w\" (UniqueName: \"kubernetes.io/projected/5b2d9fda-bef0-4a73-84aa-c793181811d9-kube-api-access-8t42w\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:04 crc kubenswrapper[4809]: I1205 11:30:04.885144 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" path="/var/lib/kubelet/pods/de639392-c7d7-4b7f-b894-0fb5d32f00a2/volumes" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.194649 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22pc6" event={"ID":"5b2d9fda-bef0-4a73-84aa-c793181811d9","Type":"ContainerDied","Data":"e5212bd4f5c7fa3d8ac934217fc6c4f33206fa64473d0f18adb7e9037197d43a"} Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.194685 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5212bd4f5c7fa3d8ac934217fc6c4f33206fa64473d0f18adb7e9037197d43a" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.194752 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22pc6" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.506704 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l"] Dec 05 11:30:05 crc kubenswrapper[4809]: E1205 11:30:05.507778 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerName="dnsmasq-dns" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.507913 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerName="dnsmasq-dns" Dec 05 11:30:05 crc kubenswrapper[4809]: E1205 11:30:05.508021 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ee7bf0-2a51-4926-b47d-c06c792c1fe9" containerName="collect-profiles" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.508084 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ee7bf0-2a51-4926-b47d-c06c792c1fe9" containerName="collect-profiles" Dec 05 11:30:05 crc kubenswrapper[4809]: E1205 11:30:05.508163 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2d9fda-bef0-4a73-84aa-c793181811d9" containerName="keystone-db-sync" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.508226 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2d9fda-bef0-4a73-84aa-c793181811d9" containerName="keystone-db-sync" Dec 05 11:30:05 crc kubenswrapper[4809]: E1205 11:30:05.508295 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerName="init" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.508371 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerName="init" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.508685 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="de639392-c7d7-4b7f-b894-0fb5d32f00a2" containerName="dnsmasq-dns" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.508772 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2d9fda-bef0-4a73-84aa-c793181811d9" containerName="keystone-db-sync" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.508850 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ee7bf0-2a51-4926-b47d-c06c792c1fe9" containerName="collect-profiles" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.510473 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.520286 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fp8h9"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.522468 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.535086 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dnjt6" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.535572 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.535766 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.535920 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.536111 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.539666 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.552849 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fp8h9"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-credential-keys\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587781 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-fernet-keys\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587801 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-config\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-config-data\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587946 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-scripts\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.587974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgssz\" (UniqueName: \"kubernetes.io/projected/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-kube-api-access-fgssz\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.588007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b22qf\" (UniqueName: \"kubernetes.io/projected/d1779fdc-f365-455d-a666-22002b3851a7-kube-api-access-b22qf\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.588044 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-combined-ca-bundle\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.588063 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.588103 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.698678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.699706 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.699824 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.699892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-config\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.699966 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-config-data\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-scripts\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgssz\" (UniqueName: \"kubernetes.io/projected/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-kube-api-access-fgssz\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b22qf\" (UniqueName: \"kubernetes.io/projected/d1779fdc-f365-455d-a666-22002b3851a7-kube-api-access-b22qf\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700235 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-combined-ca-bundle\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700277 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700328 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700484 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-credential-keys\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.700557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-fernet-keys\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.702269 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-config\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.702911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.703659 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.705228 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.710002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-scripts\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.710372 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-fernet-keys\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.714884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-combined-ca-bundle\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.722975 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-credential-keys\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.723217 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-config-data\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.736572 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgssz\" (UniqueName: \"kubernetes.io/projected/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-kube-api-access-fgssz\") pod \"dnsmasq-dns-5c5cc7c5ff-t5v9l\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.752141 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b22qf\" (UniqueName: \"kubernetes.io/projected/d1779fdc-f365-455d-a666-22002b3851a7-kube-api-access-b22qf\") pod \"keystone-bootstrap-fp8h9\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.828271 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-knlr7"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.831951 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.838237 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.838579 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.838781 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q2l2q" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.849411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.854666 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.862418 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-9mn2t"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.867343 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.883299 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.883717 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-f998k" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.883860 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.905446 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp9tm\" (UniqueName: \"kubernetes.io/projected/eb75204b-f255-48c4-bc67-54bd3601c6a9-kube-api-access-zp9tm\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.905572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-combined-ca-bundle\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.905613 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-scripts\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.905683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb75204b-f255-48c4-bc67-54bd3601c6a9-etc-machine-id\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.905709 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-db-sync-config-data\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.914204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-config-data\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.918205 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-knlr7"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.949032 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.951747 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.965243 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.967042 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9mn2t"] Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.979409 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 11:30:05 crc kubenswrapper[4809]: I1205 11:30:05.998133 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.017890 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrqvl\" (UniqueName: \"kubernetes.io/projected/cc0e50e9-996a-4a27-aa6f-97bba30a0427-kube-api-access-lrqvl\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.017944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-config\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.017991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.018015 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb75204b-f255-48c4-bc67-54bd3601c6a9-etc-machine-id\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.018034 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-db-sync-config-data\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.018093 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-combined-ca-bundle\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.018155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-config-data\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.018191 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-config-data\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.018223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.018268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-run-httpd\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.020975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-log-httpd\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.021088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztjkl\" (UniqueName: \"kubernetes.io/projected/09271e87-1a57-43fb-81ca-8e6221dc81e0-kube-api-access-ztjkl\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.021120 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-scripts\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.021167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp9tm\" (UniqueName: \"kubernetes.io/projected/eb75204b-f255-48c4-bc67-54bd3601c6a9-kube-api-access-zp9tm\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.033791 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-combined-ca-bundle\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.033899 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-scripts\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.035595 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb75204b-f255-48c4-bc67-54bd3601c6a9-etc-machine-id\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.049407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-combined-ca-bundle\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.050258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-db-sync-config-data\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.054400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-scripts\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.056316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-config-data\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.082858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp9tm\" (UniqueName: \"kubernetes.io/projected/eb75204b-f255-48c4-bc67-54bd3601c6a9-kube-api-access-zp9tm\") pod \"cinder-db-sync-knlr7\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.123651 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ddjc8"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.133146 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.137864 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-combined-ca-bundle\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.137993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-config-data\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138310 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-run-httpd\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138400 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-log-httpd\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztjkl\" (UniqueName: \"kubernetes.io/projected/09271e87-1a57-43fb-81ca-8e6221dc81e0-kube-api-access-ztjkl\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138567 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-scripts\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrqvl\" (UniqueName: \"kubernetes.io/projected/cc0e50e9-996a-4a27-aa6f-97bba30a0427-kube-api-access-lrqvl\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-config\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.138830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.143884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-run-httpd\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.144408 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.147330 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.147543 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-txjnf" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.147328 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-log-httpd\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.154154 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.154959 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-config\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.175616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.176960 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-config-data\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.178208 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.179210 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-combined-ca-bundle\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.180453 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrqvl\" (UniqueName: \"kubernetes.io/projected/cc0e50e9-996a-4a27-aa6f-97bba30a0427-kube-api-access-lrqvl\") pod \"neutron-db-sync-9mn2t\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.185371 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztjkl\" (UniqueName: \"kubernetes.io/projected/09271e87-1a57-43fb-81ca-8e6221dc81e0-kube-api-access-ztjkl\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.191273 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-scripts\") pod \"ceilometer-0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.194274 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ddjc8"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.212389 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.213038 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2444g"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.215777 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.218112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.219422 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-n5dxb" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.219776 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.229720 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2444g"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.242961 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98ab6cf6-a8a1-43d5-8c37-8eddae072869-logs\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.243141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmwk\" (UniqueName: \"kubernetes.io/projected/98ab6cf6-a8a1-43d5-8c37-8eddae072869-kube-api-access-vdmwk\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.243190 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-config-data\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.243216 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-combined-ca-bundle\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.243241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-scripts\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.246071 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-4hr9q"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.247827 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.258191 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-4hr9q"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355014 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355118 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8b7s\" (UniqueName: \"kubernetes.io/projected/64f0f74c-6a4b-4549-8abb-5d497627e1bf-kube-api-access-t8b7s\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355148 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355213 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-combined-ca-bundle\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355284 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98ab6cf6-a8a1-43d5-8c37-8eddae072869-logs\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355356 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl2v9\" (UniqueName: \"kubernetes.io/projected/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-kube-api-access-fl2v9\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355389 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmwk\" (UniqueName: \"kubernetes.io/projected/98ab6cf6-a8a1-43d5-8c37-8eddae072869-kube-api-access-vdmwk\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-db-sync-config-data\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-config-data\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-combined-ca-bundle\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355545 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-scripts\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355569 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.355689 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-config\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.356289 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98ab6cf6-a8a1-43d5-8c37-8eddae072869-logs\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.361412 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.377305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmwk\" (UniqueName: \"kubernetes.io/projected/98ab6cf6-a8a1-43d5-8c37-8eddae072869-kube-api-access-vdmwk\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.377431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-config-data\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.387200 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-scripts\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.391747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-combined-ca-bundle\") pod \"placement-db-sync-ddjc8\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.457958 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8b7s\" (UniqueName: \"kubernetes.io/projected/64f0f74c-6a4b-4549-8abb-5d497627e1bf-kube-api-access-t8b7s\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458429 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-combined-ca-bundle\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl2v9\" (UniqueName: \"kubernetes.io/projected/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-kube-api-access-fl2v9\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-db-sync-config-data\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458757 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458788 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-config\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.458835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.462056 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.463115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.465446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.466749 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-config\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.467485 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.480131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.482216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-db-sync-config-data\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.483953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-combined-ca-bundle\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.497855 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8b7s\" (UniqueName: \"kubernetes.io/projected/64f0f74c-6a4b-4549-8abb-5d497627e1bf-kube-api-access-t8b7s\") pod \"barbican-db-sync-2444g\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.498232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl2v9\" (UniqueName: \"kubernetes.io/projected/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-kube-api-access-fl2v9\") pod \"dnsmasq-dns-8b5c85b87-4hr9q\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.502367 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fp8h9"] Dec 05 11:30:06 crc kubenswrapper[4809]: W1205 11:30:06.522149 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1779fdc_f365_455d_a666_22002b3851a7.slice/crio-f7ecc923f96916b4b218b502865f87f582620ac487dc232e74b515f964432362 WatchSource:0}: Error finding container f7ecc923f96916b4b218b502865f87f582620ac487dc232e74b515f964432362: Status 404 returned error can't find the container with id f7ecc923f96916b4b218b502865f87f582620ac487dc232e74b515f964432362 Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.547915 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.589651 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.670015 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.676678 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.682272 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hfvxz" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.682524 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.682913 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.683742 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.718768 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768199 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768307 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf9pj\" (UniqueName: \"kubernetes.io/projected/ee51fe72-7080-4693-80be-4e61690a6851-kube-api-access-vf9pj\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768390 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.768445 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-logs\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.821719 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.823401 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.829451 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.830220 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.857330 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.870357 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.870969 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.870996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871045 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871083 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871122 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf9pj\" (UniqueName: \"kubernetes.io/projected/ee51fe72-7080-4693-80be-4e61690a6851-kube-api-access-vf9pj\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871203 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871229 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq9tw\" (UniqueName: \"kubernetes.io/projected/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-kube-api-access-hq9tw\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871425 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871458 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.871490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-logs\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.872462 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-logs\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.872561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.872969 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.898332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-config-data\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.913413 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.914343 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf9pj\" (UniqueName: \"kubernetes.io/projected/ee51fe72-7080-4693-80be-4e61690a6851-kube-api-access-vf9pj\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.943103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-scripts\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.943348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.951083 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.966345 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq9tw\" (UniqueName: \"kubernetes.io/projected/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-kube-api-access-hq9tw\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974229 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974325 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.974388 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.978984 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.986012 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.987246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.987732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.987916 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-knlr7"] Dec 05 11:30:06 crc kubenswrapper[4809]: I1205 11:30:06.997264 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.005528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.011727 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.029564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq9tw\" (UniqueName: \"kubernetes.io/projected/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-kube-api-access-hq9tw\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.061423 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9mn2t"] Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.135011 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.145357 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.189411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.197200 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.240463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fp8h9" event={"ID":"d1779fdc-f365-455d-a666-22002b3851a7","Type":"ContainerStarted","Data":"f7ecc923f96916b4b218b502865f87f582620ac487dc232e74b515f964432362"} Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.445369 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ddjc8"] Dec 05 11:30:07 crc kubenswrapper[4809]: W1205 11:30:07.453798 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98ab6cf6_a8a1_43d5_8c37_8eddae072869.slice/crio-0ee4da898b60ef108886e9a91910b363b4f731e7954cdf478c2e8ab166dbbf1b WatchSource:0}: Error finding container 0ee4da898b60ef108886e9a91910b363b4f731e7954cdf478c2e8ab166dbbf1b: Status 404 returned error can't find the container with id 0ee4da898b60ef108886e9a91910b363b4f731e7954cdf478c2e8ab166dbbf1b Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.622558 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2444g"] Dec 05 11:30:07 crc kubenswrapper[4809]: W1205 11:30:07.629782 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64f0f74c_6a4b_4549_8abb_5d497627e1bf.slice/crio-a33323980bb5de72b98b3a0ed378804abfac195584dd0685e9c61d551d3aa1e8 WatchSource:0}: Error finding container a33323980bb5de72b98b3a0ed378804abfac195584dd0685e9c61d551d3aa1e8: Status 404 returned error can't find the container with id a33323980bb5de72b98b3a0ed378804abfac195584dd0685e9c61d551d3aa1e8 Dec 05 11:30:07 crc kubenswrapper[4809]: I1205 11:30:07.735541 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-4hr9q"] Dec 05 11:30:07 crc kubenswrapper[4809]: W1205 11:30:07.762237 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e3fe5e5_aeaa_402c_8b90_715656ecf6d9.slice/crio-b641879e2a9eea87d98a95f5f5a9278843966d4ad136f2d147572070e80e2c74 WatchSource:0}: Error finding container b641879e2a9eea87d98a95f5f5a9278843966d4ad136f2d147572070e80e2c74: Status 404 returned error can't find the container with id b641879e2a9eea87d98a95f5f5a9278843966d4ad136f2d147572070e80e2c74 Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.109797 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.142371 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.214941 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.233753 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.246515 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.275997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9mn2t" event={"ID":"cc0e50e9-996a-4a27-aa6f-97bba30a0427","Type":"ContainerStarted","Data":"699916750a9c30999954b19d9fed3c1c85febad38b4f6917f5c1b474c4e9abc2"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.290905 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-knlr7" event={"ID":"eb75204b-f255-48c4-bc67-54bd3601c6a9","Type":"ContainerStarted","Data":"aecfb9c6ed9754ed7fc6d8cdda5130e3addb4d34057aba9f979823e4f291c6de"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.313748 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d","Type":"ContainerStarted","Data":"8672b6a173876e5617c7f952bec655b262ed2ebb6387d9e9e0c26221f6e7185d"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.331923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" event={"ID":"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9","Type":"ContainerStarted","Data":"b641879e2a9eea87d98a95f5f5a9278843966d4ad136f2d147572070e80e2c74"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.333443 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee51fe72-7080-4693-80be-4e61690a6851","Type":"ContainerStarted","Data":"b42ba8764ac038c15ea300a5230f35cd58c073f921fa6c0b5ad6bf85d3d03211"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.334889 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerStarted","Data":"feb596b2dfe233f1c3c385e2dbb04a329b8b7e7937619355edc4a270ea29f784"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.336648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2444g" event={"ID":"64f0f74c-6a4b-4549-8abb-5d497627e1bf","Type":"ContainerStarted","Data":"a33323980bb5de72b98b3a0ed378804abfac195584dd0685e9c61d551d3aa1e8"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.387845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" event={"ID":"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5","Type":"ContainerStarted","Data":"c72db015f02fef5fa64fd94f423bb8af91f21042c35936b0f95ea57cdcfa848c"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.388137 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" podUID="6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" containerName="init" containerID="cri-o://bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93" gracePeriod=10 Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.396232 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ddjc8" event={"ID":"98ab6cf6-a8a1-43d5-8c37-8eddae072869","Type":"ContainerStarted","Data":"0ee4da898b60ef108886e9a91910b363b4f731e7954cdf478c2e8ab166dbbf1b"} Dec 05 11:30:08 crc kubenswrapper[4809]: I1205 11:30:08.886517 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.033253 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-sb\") pod \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.033291 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-config\") pod \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.033374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-svc\") pod \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.033396 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-swift-storage-0\") pod \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.033504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgssz\" (UniqueName: \"kubernetes.io/projected/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-kube-api-access-fgssz\") pod \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.033533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-nb\") pod \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\" (UID: \"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5\") " Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.060417 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" (UID: "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.066500 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" (UID: "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.077182 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-kube-api-access-fgssz" (OuterVolumeSpecName: "kube-api-access-fgssz") pod "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" (UID: "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5"). InnerVolumeSpecName "kube-api-access-fgssz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.082704 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" (UID: "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.108598 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-config" (OuterVolumeSpecName: "config") pod "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" (UID: "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.121057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" (UID: "6a0d5d1a-4cc7-417b-ae1e-06e345358dc5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.135947 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.135991 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.136003 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.136016 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.136028 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgssz\" (UniqueName: \"kubernetes.io/projected/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-kube-api-access-fgssz\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.136038 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.415334 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9mn2t" event={"ID":"cc0e50e9-996a-4a27-aa6f-97bba30a0427","Type":"ContainerStarted","Data":"e66ce562a684ce2f7c9a5765f371c11805fe1afdfbe8c023db95e39baf7fd55d"} Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.421150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fp8h9" event={"ID":"d1779fdc-f365-455d-a666-22002b3851a7","Type":"ContainerStarted","Data":"4124bbc4606e9f85acf01d756fb541e366fb21a27e987393909ff3ed0e222a54"} Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.428607 4809 generic.go:334] "Generic (PLEG): container finished" podID="6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" containerID="bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93" exitCode=0 Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.428750 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.428959 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" event={"ID":"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5","Type":"ContainerDied","Data":"c72db015f02fef5fa64fd94f423bb8af91f21042c35936b0f95ea57cdcfa848c"} Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.429026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l" event={"ID":"6a0d5d1a-4cc7-417b-ae1e-06e345358dc5","Type":"ContainerDied","Data":"bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93"} Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.429049 4809 scope.go:117] "RemoveContainer" containerID="bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.440076 4809 generic.go:334] "Generic (PLEG): container finished" podID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerID="d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa" exitCode=0 Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.440143 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" event={"ID":"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9","Type":"ContainerDied","Data":"d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa"} Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.446377 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-9mn2t" podStartSLOduration=4.446352909 podStartE2EDuration="4.446352909s" podCreationTimestamp="2025-12-05 11:30:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:09.429820646 +0000 UTC m=+1304.820797194" watchObservedRunningTime="2025-12-05 11:30:09.446352909 +0000 UTC m=+1304.837329467" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.455863 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fp8h9" podStartSLOduration=4.4558459280000005 podStartE2EDuration="4.455845928s" podCreationTimestamp="2025-12-05 11:30:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:09.448723221 +0000 UTC m=+1304.839699779" watchObservedRunningTime="2025-12-05 11:30:09.455845928 +0000 UTC m=+1304.846822486" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.529572 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l"] Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.551577 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-t5v9l"] Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.582274 4809 scope.go:117] "RemoveContainer" containerID="bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93" Dec 05 11:30:09 crc kubenswrapper[4809]: E1205 11:30:09.582755 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93\": container with ID starting with bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93 not found: ID does not exist" containerID="bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93" Dec 05 11:30:09 crc kubenswrapper[4809]: I1205 11:30:09.582790 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93"} err="failed to get container status \"bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93\": rpc error: code = NotFound desc = could not find container \"bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93\": container with ID starting with bef076a2d46cf23c9ca73ce6e2a0fd8f43d529634f80a3d5e959bb0b4ad14a93 not found: ID does not exist" Dec 05 11:30:10 crc kubenswrapper[4809]: I1205 11:30:10.459751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee51fe72-7080-4693-80be-4e61690a6851","Type":"ContainerStarted","Data":"7398c03f5ceb12ba1d110a8bed0f1f25977c053003db63ecb184fff7deb55747"} Dec 05 11:30:10 crc kubenswrapper[4809]: I1205 11:30:10.465542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d","Type":"ContainerStarted","Data":"f385a8b4912a91a6a7d87f375c27ec079f2532c82aa7a620b94a547fb2f91f8a"} Dec 05 11:30:10 crc kubenswrapper[4809]: I1205 11:30:10.479807 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" event={"ID":"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9","Type":"ContainerStarted","Data":"9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521"} Dec 05 11:30:10 crc kubenswrapper[4809]: I1205 11:30:10.507878 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" podStartSLOduration=4.507853034 podStartE2EDuration="4.507853034s" podCreationTimestamp="2025-12-05 11:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:10.504496186 +0000 UTC m=+1305.895472744" watchObservedRunningTime="2025-12-05 11:30:10.507853034 +0000 UTC m=+1305.898829582" Dec 05 11:30:10 crc kubenswrapper[4809]: I1205 11:30:10.891095 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" path="/var/lib/kubelet/pods/6a0d5d1a-4cc7-417b-ae1e-06e345358dc5/volumes" Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.493283 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee51fe72-7080-4693-80be-4e61690a6851","Type":"ContainerStarted","Data":"fa718d7d81f951cac4bc90c2a5e357ae0cada48668e29e628b6382eb3a41da71"} Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.493434 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-log" containerID="cri-o://7398c03f5ceb12ba1d110a8bed0f1f25977c053003db63ecb184fff7deb55747" gracePeriod=30 Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.494017 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-httpd" containerID="cri-o://fa718d7d81f951cac4bc90c2a5e357ae0cada48668e29e628b6382eb3a41da71" gracePeriod=30 Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.505556 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-log" containerID="cri-o://f385a8b4912a91a6a7d87f375c27ec079f2532c82aa7a620b94a547fb2f91f8a" gracePeriod=30 Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.505664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d","Type":"ContainerStarted","Data":"400a18979c214ed616801ebf386b6f24c8ffa3b41730bfcbdf1fd5b17e167cb0"} Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.505688 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.505729 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-httpd" containerID="cri-o://400a18979c214ed616801ebf386b6f24c8ffa3b41730bfcbdf1fd5b17e167cb0" gracePeriod=30 Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.527892 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.527861869 podStartE2EDuration="6.527861869s" podCreationTimestamp="2025-12-05 11:30:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:11.514994162 +0000 UTC m=+1306.905970720" watchObservedRunningTime="2025-12-05 11:30:11.527861869 +0000 UTC m=+1306.918838427" Dec 05 11:30:11 crc kubenswrapper[4809]: I1205 11:30:11.547154 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.547129995 podStartE2EDuration="6.547129995s" podCreationTimestamp="2025-12-05 11:30:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:11.539585947 +0000 UTC m=+1306.930562505" watchObservedRunningTime="2025-12-05 11:30:11.547129995 +0000 UTC m=+1306.938106553" Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.518421 4809 generic.go:334] "Generic (PLEG): container finished" podID="ee51fe72-7080-4693-80be-4e61690a6851" containerID="fa718d7d81f951cac4bc90c2a5e357ae0cada48668e29e628b6382eb3a41da71" exitCode=0 Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.518786 4809 generic.go:334] "Generic (PLEG): container finished" podID="ee51fe72-7080-4693-80be-4e61690a6851" containerID="7398c03f5ceb12ba1d110a8bed0f1f25977c053003db63ecb184fff7deb55747" exitCode=143 Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.518497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee51fe72-7080-4693-80be-4e61690a6851","Type":"ContainerDied","Data":"fa718d7d81f951cac4bc90c2a5e357ae0cada48668e29e628b6382eb3a41da71"} Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.518871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee51fe72-7080-4693-80be-4e61690a6851","Type":"ContainerDied","Data":"7398c03f5ceb12ba1d110a8bed0f1f25977c053003db63ecb184fff7deb55747"} Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.521590 4809 generic.go:334] "Generic (PLEG): container finished" podID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerID="400a18979c214ed616801ebf386b6f24c8ffa3b41730bfcbdf1fd5b17e167cb0" exitCode=0 Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.521616 4809 generic.go:334] "Generic (PLEG): container finished" podID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerID="f385a8b4912a91a6a7d87f375c27ec079f2532c82aa7a620b94a547fb2f91f8a" exitCode=143 Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.521680 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d","Type":"ContainerDied","Data":"400a18979c214ed616801ebf386b6f24c8ffa3b41730bfcbdf1fd5b17e167cb0"} Dec 05 11:30:12 crc kubenswrapper[4809]: I1205 11:30:12.521738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d","Type":"ContainerDied","Data":"f385a8b4912a91a6a7d87f375c27ec079f2532c82aa7a620b94a547fb2f91f8a"} Dec 05 11:30:13 crc kubenswrapper[4809]: I1205 11:30:13.539905 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1779fdc-f365-455d-a666-22002b3851a7" containerID="4124bbc4606e9f85acf01d756fb541e366fb21a27e987393909ff3ed0e222a54" exitCode=0 Dec 05 11:30:13 crc kubenswrapper[4809]: I1205 11:30:13.540243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fp8h9" event={"ID":"d1779fdc-f365-455d-a666-22002b3851a7","Type":"ContainerDied","Data":"4124bbc4606e9f85acf01d756fb541e366fb21a27e987393909ff3ed0e222a54"} Dec 05 11:30:14 crc kubenswrapper[4809]: I1205 11:30:14.046873 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:30:14 crc kubenswrapper[4809]: I1205 11:30:14.047234 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:30:16 crc kubenswrapper[4809]: I1205 11:30:16.591783 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:16 crc kubenswrapper[4809]: I1205 11:30:16.713106 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jmvtl"] Dec 05 11:30:16 crc kubenswrapper[4809]: I1205 11:30:16.713327 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="dnsmasq-dns" containerID="cri-o://f615d4be35da2629c2d18907ba7e6166ab3ef4c48f963d7e1f9a7707bd8f6399" gracePeriod=10 Dec 05 11:30:17 crc kubenswrapper[4809]: I1205 11:30:17.603303 4809 generic.go:334] "Generic (PLEG): container finished" podID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerID="f615d4be35da2629c2d18907ba7e6166ab3ef4c48f963d7e1f9a7707bd8f6399" exitCode=0 Dec 05 11:30:17 crc kubenswrapper[4809]: I1205 11:30:17.603345 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" event={"ID":"b8721dce-921b-4ac4-ae84-94815ae99e67","Type":"ContainerDied","Data":"f615d4be35da2629c2d18907ba7e6166ab3ef4c48f963d7e1f9a7707bd8f6399"} Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.128270 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.140797 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.141396 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b22qf\" (UniqueName: \"kubernetes.io/projected/d1779fdc-f365-455d-a666-22002b3851a7-kube-api-access-b22qf\") pod \"d1779fdc-f365-455d-a666-22002b3851a7\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.141557 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-combined-ca-bundle\") pod \"d1779fdc-f365-455d-a666-22002b3851a7\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.141734 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-config-data\") pod \"d1779fdc-f365-455d-a666-22002b3851a7\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.141798 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-scripts\") pod \"d1779fdc-f365-455d-a666-22002b3851a7\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.141849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-credential-keys\") pod \"d1779fdc-f365-455d-a666-22002b3851a7\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.141926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-fernet-keys\") pod \"d1779fdc-f365-455d-a666-22002b3851a7\" (UID: \"d1779fdc-f365-455d-a666-22002b3851a7\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.152015 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-scripts" (OuterVolumeSpecName: "scripts") pod "d1779fdc-f365-455d-a666-22002b3851a7" (UID: "d1779fdc-f365-455d-a666-22002b3851a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.153527 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1779fdc-f365-455d-a666-22002b3851a7-kube-api-access-b22qf" (OuterVolumeSpecName: "kube-api-access-b22qf") pod "d1779fdc-f365-455d-a666-22002b3851a7" (UID: "d1779fdc-f365-455d-a666-22002b3851a7"). InnerVolumeSpecName "kube-api-access-b22qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.154651 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.157207 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d1779fdc-f365-455d-a666-22002b3851a7" (UID: "d1779fdc-f365-455d-a666-22002b3851a7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.158774 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d1779fdc-f365-455d-a666-22002b3851a7" (UID: "d1779fdc-f365-455d-a666-22002b3851a7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.219121 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-config-data" (OuterVolumeSpecName: "config-data") pod "d1779fdc-f365-455d-a666-22002b3851a7" (UID: "d1779fdc-f365-455d-a666-22002b3851a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.240903 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1779fdc-f365-455d-a666-22002b3851a7" (UID: "d1779fdc-f365-455d-a666-22002b3851a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.244563 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.244752 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-config-data\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.244909 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-combined-ca-bundle\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245029 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-public-tls-certs\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245242 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-logs\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-httpd-run\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-logs\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-scripts\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245501 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-logs" (OuterVolumeSpecName: "logs") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245705 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245784 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq9tw\" (UniqueName: \"kubernetes.io/projected/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-kube-api-access-hq9tw\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-logs" (OuterVolumeSpecName: "logs") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245932 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-httpd-run\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.245974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf9pj\" (UniqueName: \"kubernetes.io/projected/ee51fe72-7080-4693-80be-4e61690a6851-kube-api-access-vf9pj\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246039 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-scripts\") pod \"ee51fe72-7080-4693-80be-4e61690a6851\" (UID: \"ee51fe72-7080-4693-80be-4e61690a6851\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-combined-ca-bundle\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-config-data\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246171 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-internal-tls-certs\") pod \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\" (UID: \"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d\") " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246865 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246891 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246901 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246914 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246925 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246937 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246947 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b22qf\" (UniqueName: \"kubernetes.io/projected/d1779fdc-f365-455d-a666-22002b3851a7-kube-api-access-b22qf\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246957 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee51fe72-7080-4693-80be-4e61690a6851-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.246968 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1779fdc-f365-455d-a666-22002b3851a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.247660 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.247726 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.249303 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.252342 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-scripts" (OuterVolumeSpecName: "scripts") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.255719 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-scripts" (OuterVolumeSpecName: "scripts") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.258867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee51fe72-7080-4693-80be-4e61690a6851-kube-api-access-vf9pj" (OuterVolumeSpecName: "kube-api-access-vf9pj") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "kube-api-access-vf9pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.267551 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-kube-api-access-hq9tw" (OuterVolumeSpecName: "kube-api-access-hq9tw") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "kube-api-access-hq9tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.282361 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.297238 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.310779 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-config-data" (OuterVolumeSpecName: "config-data") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.320738 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-config-data" (OuterVolumeSpecName: "config-data") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.331911 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" (UID: "f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.339866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ee51fe72-7080-4693-80be-4e61690a6851" (UID: "ee51fe72-7080-4693-80be-4e61690a6851"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.348941 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq9tw\" (UniqueName: \"kubernetes.io/projected/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-kube-api-access-hq9tw\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.348979 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.348989 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf9pj\" (UniqueName: \"kubernetes.io/projected/ee51fe72-7080-4693-80be-4e61690a6851-kube-api-access-vf9pj\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.348997 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349009 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349018 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349027 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349061 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349071 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349081 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349094 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349103 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee51fe72-7080-4693-80be-4e61690a6851-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.349111 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.367497 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.369813 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.451112 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.451168 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.622910 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fp8h9" event={"ID":"d1779fdc-f365-455d-a666-22002b3851a7","Type":"ContainerDied","Data":"f7ecc923f96916b4b218b502865f87f582620ac487dc232e74b515f964432362"} Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.622978 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7ecc923f96916b4b218b502865f87f582620ac487dc232e74b515f964432362" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.622999 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fp8h9" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.625681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ee51fe72-7080-4693-80be-4e61690a6851","Type":"ContainerDied","Data":"b42ba8764ac038c15ea300a5230f35cd58c073f921fa6c0b5ad6bf85d3d03211"} Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.625771 4809 scope.go:117] "RemoveContainer" containerID="fa718d7d81f951cac4bc90c2a5e357ae0cada48668e29e628b6382eb3a41da71" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.626077 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.633591 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d","Type":"ContainerDied","Data":"8672b6a173876e5617c7f952bec655b262ed2ebb6387d9e9e0c26221f6e7185d"} Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.633741 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.704573 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.711678 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.732075 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.740982 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748245 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: E1205 11:30:18.748608 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-httpd" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748646 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-httpd" Dec 05 11:30:18 crc kubenswrapper[4809]: E1205 11:30:18.748662 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1779fdc-f365-455d-a666-22002b3851a7" containerName="keystone-bootstrap" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748671 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1779fdc-f365-455d-a666-22002b3851a7" containerName="keystone-bootstrap" Dec 05 11:30:18 crc kubenswrapper[4809]: E1205 11:30:18.748683 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-log" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748688 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-log" Dec 05 11:30:18 crc kubenswrapper[4809]: E1205 11:30:18.748698 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-httpd" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748703 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-httpd" Dec 05 11:30:18 crc kubenswrapper[4809]: E1205 11:30:18.748715 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" containerName="init" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748720 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" containerName="init" Dec 05 11:30:18 crc kubenswrapper[4809]: E1205 11:30:18.748742 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-log" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748750 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-log" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748909 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-httpd" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748920 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1779fdc-f365-455d-a666-22002b3851a7" containerName="keystone-bootstrap" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748928 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee51fe72-7080-4693-80be-4e61690a6851" containerName="glance-log" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748940 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-log" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748947 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0d5d1a-4cc7-417b-ae1e-06e345358dc5" containerName="init" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.748964 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" containerName="glance-httpd" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.749975 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.763649 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.765770 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hfvxz" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.766126 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.766246 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.766308 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.770471 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.775130 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.775577 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.782140 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.796944 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.859679 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.860618 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.860729 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdx96\" (UniqueName: \"kubernetes.io/projected/2c4570ff-b46c-48e1-9aff-1b460ba27581-kube-api-access-tdx96\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.860807 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.860903 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.860937 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-logs\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.860969 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.861007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.885752 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee51fe72-7080-4693-80be-4e61690a6851" path="/var/lib/kubelet/pods/ee51fe72-7080-4693-80be-4e61690a6851/volumes" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.886610 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d" path="/var/lib/kubelet/pods/f7cd0d2c-fd38-490f-b0d0-ff23be62ed6d/volumes" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.964248 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.964334 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58s8m\" (UniqueName: \"kubernetes.io/projected/157daa4d-2576-408d-a86d-c9c6e9c162e1-kube-api-access-58s8m\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.964382 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.964702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.964787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.964831 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdx96\" (UniqueName: \"kubernetes.io/projected/2c4570ff-b46c-48e1-9aff-1b460ba27581-kube-api-access-tdx96\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.964988 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965060 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965258 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-logs\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965306 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965334 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-logs\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965424 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-config-data\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.965456 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-scripts\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.971484 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.971668 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-logs\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.972141 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.973253 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.973531 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.977901 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.978558 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:18 crc kubenswrapper[4809]: I1205 11:30:18.997203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdx96\" (UniqueName: \"kubernetes.io/projected/2c4570ff-b46c-48e1-9aff-1b460ba27581-kube-api-access-tdx96\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.012029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.067287 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.067860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-logs\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.067898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.067945 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-config-data\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.067981 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-scripts\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.068018 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58s8m\" (UniqueName: \"kubernetes.io/projected/157daa4d-2576-408d-a86d-c9c6e9c162e1-kube-api-access-58s8m\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.068520 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-logs\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.068521 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.068954 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.069007 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.069428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.074170 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.074450 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-scripts\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.082449 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-config-data\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.083220 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.092313 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58s8m\" (UniqueName: \"kubernetes.io/projected/157daa4d-2576-408d-a86d-c9c6e9c162e1-kube-api-access-58s8m\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.104081 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.124316 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.242631 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fp8h9"] Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.253949 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fp8h9"] Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.314992 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-77mnx"] Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.318566 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.322096 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.322299 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.322484 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dnjt6" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.323122 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.323882 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.330817 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-77mnx"] Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.383742 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-combined-ca-bundle\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.383821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-fernet-keys\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.383849 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnm9d\" (UniqueName: \"kubernetes.io/projected/34702fbf-a89c-46ed-9546-bc873caee32e-kube-api-access-fnm9d\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.383868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-scripts\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.383916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-config-data\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.383937 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-credential-keys\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.411401 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.491076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-fernet-keys\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.491129 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnm9d\" (UniqueName: \"kubernetes.io/projected/34702fbf-a89c-46ed-9546-bc873caee32e-kube-api-access-fnm9d\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.491149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-scripts\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.491204 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-config-data\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.491227 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-credential-keys\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.491286 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-combined-ca-bundle\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.495116 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-scripts\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.495953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-combined-ca-bundle\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.497275 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-fernet-keys\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.498242 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-credential-keys\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.500191 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-config-data\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.514372 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnm9d\" (UniqueName: \"kubernetes.io/projected/34702fbf-a89c-46ed-9546-bc873caee32e-kube-api-access-fnm9d\") pod \"keystone-bootstrap-77mnx\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:19 crc kubenswrapper[4809]: I1205 11:30:19.648726 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:20 crc kubenswrapper[4809]: I1205 11:30:20.883602 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1779fdc-f365-455d-a666-22002b3851a7" path="/var/lib/kubelet/pods/d1779fdc-f365-455d-a666-22002b3851a7/volumes" Dec 05 11:30:21 crc kubenswrapper[4809]: I1205 11:30:21.673525 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Dec 05 11:30:26 crc kubenswrapper[4809]: I1205 11:30:26.673653 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Dec 05 11:30:27 crc kubenswrapper[4809]: E1205 11:30:27.481881 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 05 11:30:27 crc kubenswrapper[4809]: E1205 11:30:27.482344 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t8b7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-2444g_openstack(64f0f74c-6a4b-4549-8abb-5d497627e1bf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:30:27 crc kubenswrapper[4809]: E1205 11:30:27.484572 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-2444g" podUID="64f0f74c-6a4b-4549-8abb-5d497627e1bf" Dec 05 11:30:27 crc kubenswrapper[4809]: E1205 11:30:27.740970 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-2444g" podUID="64f0f74c-6a4b-4549-8abb-5d497627e1bf" Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.615783 4809 scope.go:117] "RemoveContainer" containerID="7398c03f5ceb12ba1d110a8bed0f1f25977c053003db63ecb184fff7deb55747" Dec 05 11:30:28 crc kubenswrapper[4809]: E1205 11:30:28.633695 4809 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 05 11:30:28 crc kubenswrapper[4809]: E1205 11:30:28.633903 4809 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zp9tm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-knlr7_openstack(eb75204b-f255-48c4-bc67-54bd3601c6a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 11:30:28 crc kubenswrapper[4809]: E1205 11:30:28.635378 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-knlr7" podUID="eb75204b-f255-48c4-bc67-54bd3601c6a9" Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.763044 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" event={"ID":"b8721dce-921b-4ac4-ae84-94815ae99e67","Type":"ContainerDied","Data":"31cbd4b24d9a8faa6963097c0cae290acbb66f065ef52da499c7bd9fd7337906"} Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.763122 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31cbd4b24d9a8faa6963097c0cae290acbb66f065ef52da499c7bd9fd7337906" Dec 05 11:30:28 crc kubenswrapper[4809]: E1205 11:30:28.771997 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-knlr7" podUID="eb75204b-f255-48c4-bc67-54bd3601c6a9" Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.847541 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.849539 4809 scope.go:117] "RemoveContainer" containerID="400a18979c214ed616801ebf386b6f24c8ffa3b41730bfcbdf1fd5b17e167cb0" Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.917508 4809 scope.go:117] "RemoveContainer" containerID="f385a8b4912a91a6a7d87f375c27ec079f2532c82aa7a620b94a547fb2f91f8a" Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.992509 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-svc\") pod \"b8721dce-921b-4ac4-ae84-94815ae99e67\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.992659 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-nb\") pod \"b8721dce-921b-4ac4-ae84-94815ae99e67\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.992740 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-swift-storage-0\") pod \"b8721dce-921b-4ac4-ae84-94815ae99e67\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.993531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-sb\") pod \"b8721dce-921b-4ac4-ae84-94815ae99e67\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.993595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-config\") pod \"b8721dce-921b-4ac4-ae84-94815ae99e67\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " Dec 05 11:30:28 crc kubenswrapper[4809]: I1205 11:30:28.993726 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7wlr\" (UniqueName: \"kubernetes.io/projected/b8721dce-921b-4ac4-ae84-94815ae99e67-kube-api-access-f7wlr\") pod \"b8721dce-921b-4ac4-ae84-94815ae99e67\" (UID: \"b8721dce-921b-4ac4-ae84-94815ae99e67\") " Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.008942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8721dce-921b-4ac4-ae84-94815ae99e67-kube-api-access-f7wlr" (OuterVolumeSpecName: "kube-api-access-f7wlr") pod "b8721dce-921b-4ac4-ae84-94815ae99e67" (UID: "b8721dce-921b-4ac4-ae84-94815ae99e67"). InnerVolumeSpecName "kube-api-access-f7wlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.048084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b8721dce-921b-4ac4-ae84-94815ae99e67" (UID: "b8721dce-921b-4ac4-ae84-94815ae99e67"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.053488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b8721dce-921b-4ac4-ae84-94815ae99e67" (UID: "b8721dce-921b-4ac4-ae84-94815ae99e67"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.060778 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b8721dce-921b-4ac4-ae84-94815ae99e67" (UID: "b8721dce-921b-4ac4-ae84-94815ae99e67"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.071945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b8721dce-921b-4ac4-ae84-94815ae99e67" (UID: "b8721dce-921b-4ac4-ae84-94815ae99e67"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.076054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-config" (OuterVolumeSpecName: "config") pod "b8721dce-921b-4ac4-ae84-94815ae99e67" (UID: "b8721dce-921b-4ac4-ae84-94815ae99e67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.095020 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7wlr\" (UniqueName: \"kubernetes.io/projected/b8721dce-921b-4ac4-ae84-94815ae99e67-kube-api-access-f7wlr\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.095060 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.095072 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.095080 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.095088 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.095096 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8721dce-921b-4ac4-ae84-94815ae99e67-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.227423 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-77mnx"] Dec 05 11:30:29 crc kubenswrapper[4809]: W1205 11:30:29.233447 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34702fbf_a89c_46ed_9546_bc873caee32e.slice/crio-f805ca6a121a1749dc6ffb73a6e2231313bdac33e23f2a403d0a361d785a1908 WatchSource:0}: Error finding container f805ca6a121a1749dc6ffb73a6e2231313bdac33e23f2a403d0a361d785a1908: Status 404 returned error can't find the container with id f805ca6a121a1749dc6ffb73a6e2231313bdac33e23f2a403d0a361d785a1908 Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.240848 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.330434 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.418090 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.786935 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"157daa4d-2576-408d-a86d-c9c6e9c162e1","Type":"ContainerStarted","Data":"68611642e279ddb57dc391443fe6cccefdb2744dd861c8937d91fb5911a8eaaf"} Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.789013 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ddjc8" event={"ID":"98ab6cf6-a8a1-43d5-8c37-8eddae072869","Type":"ContainerStarted","Data":"1ad92427b11dd57c6028550f7cc630d7917be3d0d72e445aa75f2d9dd8a6ac7b"} Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.799157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerStarted","Data":"f3a2ecf14c34773455555834aec2896c579cd69f3d3d9bbe14f6ff7e9b81d54a"} Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.802663 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2c4570ff-b46c-48e1-9aff-1b460ba27581","Type":"ContainerStarted","Data":"832e6b0b010a706b9054e10290dfc7e15e925ecc0a20eb3c6ca0ed3c872b50ef"} Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.810358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-77mnx" event={"ID":"34702fbf-a89c-46ed-9546-bc873caee32e","Type":"ContainerStarted","Data":"62e9456b1210de31be32cf45c13c879bc91ee026ca8abe1d1406a82b2c2bc9fb"} Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.810409 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-77mnx" event={"ID":"34702fbf-a89c-46ed-9546-bc873caee32e","Type":"ContainerStarted","Data":"f805ca6a121a1749dc6ffb73a6e2231313bdac33e23f2a403d0a361d785a1908"} Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.814974 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-jmvtl" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.834276 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ddjc8" podStartSLOduration=3.833295572 podStartE2EDuration="23.834249335s" podCreationTimestamp="2025-12-05 11:30:06 +0000 UTC" firstStartedPulling="2025-12-05 11:30:07.466043744 +0000 UTC m=+1302.857020302" lastFinishedPulling="2025-12-05 11:30:27.466997507 +0000 UTC m=+1322.857974065" observedRunningTime="2025-12-05 11:30:29.807446272 +0000 UTC m=+1325.198422850" watchObservedRunningTime="2025-12-05 11:30:29.834249335 +0000 UTC m=+1325.225225903" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.842272 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-77mnx" podStartSLOduration=10.842259265 podStartE2EDuration="10.842259265s" podCreationTimestamp="2025-12-05 11:30:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:29.83330378 +0000 UTC m=+1325.224280338" watchObservedRunningTime="2025-12-05 11:30:29.842259265 +0000 UTC m=+1325.233235823" Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.895677 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jmvtl"] Dec 05 11:30:29 crc kubenswrapper[4809]: I1205 11:30:29.914675 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-jmvtl"] Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.832737 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerStarted","Data":"4f69ac2333a4fc25a147531668dcd1cda35a5d1b56826ba9ae94f1a6d4e121da"} Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.835241 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2c4570ff-b46c-48e1-9aff-1b460ba27581","Type":"ContainerStarted","Data":"ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035"} Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.835263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2c4570ff-b46c-48e1-9aff-1b460ba27581","Type":"ContainerStarted","Data":"596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99"} Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.839437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"157daa4d-2576-408d-a86d-c9c6e9c162e1","Type":"ContainerStarted","Data":"fa1e16959dbb8870e1c65a26b85f2c8b1bd0fa38bd4406d6708a3c594ed2f0c5"} Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.839463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"157daa4d-2576-408d-a86d-c9c6e9c162e1","Type":"ContainerStarted","Data":"a7397be14fac2da1c585e4bf469b3f7e487a264784a94fc89ff5aa44fb1cf14a"} Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.862812 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.862790545 podStartE2EDuration="12.862790545s" podCreationTimestamp="2025-12-05 11:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:30.852625248 +0000 UTC m=+1326.243601806" watchObservedRunningTime="2025-12-05 11:30:30.862790545 +0000 UTC m=+1326.253767093" Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.883044 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" path="/var/lib/kubelet/pods/b8721dce-921b-4ac4-ae84-94815ae99e67/volumes" Dec 05 11:30:30 crc kubenswrapper[4809]: I1205 11:30:30.900480 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=12.900456763 podStartE2EDuration="12.900456763s" podCreationTimestamp="2025-12-05 11:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:30.890062261 +0000 UTC m=+1326.281038819" watchObservedRunningTime="2025-12-05 11:30:30.900456763 +0000 UTC m=+1326.291433321" Dec 05 11:30:34 crc kubenswrapper[4809]: I1205 11:30:34.882663 4809 generic.go:334] "Generic (PLEG): container finished" podID="98ab6cf6-a8a1-43d5-8c37-8eddae072869" containerID="1ad92427b11dd57c6028550f7cc630d7917be3d0d72e445aa75f2d9dd8a6ac7b" exitCode=0 Dec 05 11:30:34 crc kubenswrapper[4809]: I1205 11:30:34.885743 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ddjc8" event={"ID":"98ab6cf6-a8a1-43d5-8c37-8eddae072869","Type":"ContainerDied","Data":"1ad92427b11dd57c6028550f7cc630d7917be3d0d72e445aa75f2d9dd8a6ac7b"} Dec 05 11:30:35 crc kubenswrapper[4809]: I1205 11:30:35.895308 4809 generic.go:334] "Generic (PLEG): container finished" podID="34702fbf-a89c-46ed-9546-bc873caee32e" containerID="62e9456b1210de31be32cf45c13c879bc91ee026ca8abe1d1406a82b2c2bc9fb" exitCode=0 Dec 05 11:30:35 crc kubenswrapper[4809]: I1205 11:30:35.895473 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-77mnx" event={"ID":"34702fbf-a89c-46ed-9546-bc873caee32e","Type":"ContainerDied","Data":"62e9456b1210de31be32cf45c13c879bc91ee026ca8abe1d1406a82b2c2bc9fb"} Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.219503 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.366243 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-combined-ca-bundle\") pod \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.366290 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdmwk\" (UniqueName: \"kubernetes.io/projected/98ab6cf6-a8a1-43d5-8c37-8eddae072869-kube-api-access-vdmwk\") pod \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.366368 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-scripts\") pod \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.366410 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98ab6cf6-a8a1-43d5-8c37-8eddae072869-logs\") pod \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.367048 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-config-data\") pod \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\" (UID: \"98ab6cf6-a8a1-43d5-8c37-8eddae072869\") " Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.367297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98ab6cf6-a8a1-43d5-8c37-8eddae072869-logs" (OuterVolumeSpecName: "logs") pod "98ab6cf6-a8a1-43d5-8c37-8eddae072869" (UID: "98ab6cf6-a8a1-43d5-8c37-8eddae072869"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.367455 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98ab6cf6-a8a1-43d5-8c37-8eddae072869-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.370844 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ab6cf6-a8a1-43d5-8c37-8eddae072869-kube-api-access-vdmwk" (OuterVolumeSpecName: "kube-api-access-vdmwk") pod "98ab6cf6-a8a1-43d5-8c37-8eddae072869" (UID: "98ab6cf6-a8a1-43d5-8c37-8eddae072869"). InnerVolumeSpecName "kube-api-access-vdmwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.370940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-scripts" (OuterVolumeSpecName: "scripts") pod "98ab6cf6-a8a1-43d5-8c37-8eddae072869" (UID: "98ab6cf6-a8a1-43d5-8c37-8eddae072869"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.394949 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98ab6cf6-a8a1-43d5-8c37-8eddae072869" (UID: "98ab6cf6-a8a1-43d5-8c37-8eddae072869"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.401941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-config-data" (OuterVolumeSpecName: "config-data") pod "98ab6cf6-a8a1-43d5-8c37-8eddae072869" (UID: "98ab6cf6-a8a1-43d5-8c37-8eddae072869"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.468579 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.468610 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.468622 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ab6cf6-a8a1-43d5-8c37-8eddae072869-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.468661 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdmwk\" (UniqueName: \"kubernetes.io/projected/98ab6cf6-a8a1-43d5-8c37-8eddae072869-kube-api-access-vdmwk\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.911461 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ddjc8" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.911414 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ddjc8" event={"ID":"98ab6cf6-a8a1-43d5-8c37-8eddae072869","Type":"ContainerDied","Data":"0ee4da898b60ef108886e9a91910b363b4f731e7954cdf478c2e8ab166dbbf1b"} Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.912210 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ee4da898b60ef108886e9a91910b363b4f731e7954cdf478c2e8ab166dbbf1b" Dec 05 11:30:36 crc kubenswrapper[4809]: I1205 11:30:36.920968 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerStarted","Data":"be78e9bdf81a9605efc222f8084146e4de3fee13231f5195827f2c22728202ee"} Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.019825 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-56b88bc7b6-8ckz6"] Dec 05 11:30:37 crc kubenswrapper[4809]: E1205 11:30:37.020397 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="init" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.020426 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="init" Dec 05 11:30:37 crc kubenswrapper[4809]: E1205 11:30:37.020450 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="dnsmasq-dns" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.020457 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="dnsmasq-dns" Dec 05 11:30:37 crc kubenswrapper[4809]: E1205 11:30:37.020471 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ab6cf6-a8a1-43d5-8c37-8eddae072869" containerName="placement-db-sync" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.020480 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ab6cf6-a8a1-43d5-8c37-8eddae072869" containerName="placement-db-sync" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.020731 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ab6cf6-a8a1-43d5-8c37-8eddae072869" containerName="placement-db-sync" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.020753 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8721dce-921b-4ac4-ae84-94815ae99e67" containerName="dnsmasq-dns" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.021981 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.027391 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.027709 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-txjnf" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.028084 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.029168 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.045788 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.055107 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-56b88bc7b6-8ckz6"] Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.183584 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-scripts\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.183705 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-public-tls-certs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.183891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-combined-ca-bundle\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.184092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-internal-tls-certs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.184142 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqjbv\" (UniqueName: \"kubernetes.io/projected/232c2be5-1e83-4fc2-b2c7-9a668d572d19-kube-api-access-zqjbv\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.184221 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232c2be5-1e83-4fc2-b2c7-9a668d572d19-logs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.184254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-config-data\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.266315 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.286339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-scripts\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.286408 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-public-tls-certs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.286482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-combined-ca-bundle\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.286547 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-internal-tls-certs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.286573 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqjbv\" (UniqueName: \"kubernetes.io/projected/232c2be5-1e83-4fc2-b2c7-9a668d572d19-kube-api-access-zqjbv\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.286612 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232c2be5-1e83-4fc2-b2c7-9a668d572d19-logs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.286654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-config-data\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.288815 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232c2be5-1e83-4fc2-b2c7-9a668d572d19-logs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.292559 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-scripts\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.293147 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-public-tls-certs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.293702 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-internal-tls-certs\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.294669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-config-data\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.297346 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-combined-ca-bundle\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.313692 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqjbv\" (UniqueName: \"kubernetes.io/projected/232c2be5-1e83-4fc2-b2c7-9a668d572d19-kube-api-access-zqjbv\") pod \"placement-56b88bc7b6-8ckz6\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.352466 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.387890 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-combined-ca-bundle\") pod \"34702fbf-a89c-46ed-9546-bc873caee32e\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.387974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-fernet-keys\") pod \"34702fbf-a89c-46ed-9546-bc873caee32e\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.388161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-credential-keys\") pod \"34702fbf-a89c-46ed-9546-bc873caee32e\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.388249 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-config-data\") pod \"34702fbf-a89c-46ed-9546-bc873caee32e\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.388423 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnm9d\" (UniqueName: \"kubernetes.io/projected/34702fbf-a89c-46ed-9546-bc873caee32e-kube-api-access-fnm9d\") pod \"34702fbf-a89c-46ed-9546-bc873caee32e\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.388501 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-scripts\") pod \"34702fbf-a89c-46ed-9546-bc873caee32e\" (UID: \"34702fbf-a89c-46ed-9546-bc873caee32e\") " Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.392658 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "34702fbf-a89c-46ed-9546-bc873caee32e" (UID: "34702fbf-a89c-46ed-9546-bc873caee32e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.392686 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "34702fbf-a89c-46ed-9546-bc873caee32e" (UID: "34702fbf-a89c-46ed-9546-bc873caee32e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.394398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34702fbf-a89c-46ed-9546-bc873caee32e-kube-api-access-fnm9d" (OuterVolumeSpecName: "kube-api-access-fnm9d") pod "34702fbf-a89c-46ed-9546-bc873caee32e" (UID: "34702fbf-a89c-46ed-9546-bc873caee32e"). InnerVolumeSpecName "kube-api-access-fnm9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.396375 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-scripts" (OuterVolumeSpecName: "scripts") pod "34702fbf-a89c-46ed-9546-bc873caee32e" (UID: "34702fbf-a89c-46ed-9546-bc873caee32e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.418877 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-config-data" (OuterVolumeSpecName: "config-data") pod "34702fbf-a89c-46ed-9546-bc873caee32e" (UID: "34702fbf-a89c-46ed-9546-bc873caee32e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.425010 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34702fbf-a89c-46ed-9546-bc873caee32e" (UID: "34702fbf-a89c-46ed-9546-bc873caee32e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.495431 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnm9d\" (UniqueName: \"kubernetes.io/projected/34702fbf-a89c-46ed-9546-bc873caee32e-kube-api-access-fnm9d\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.495472 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.495485 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.495493 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.495502 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.495510 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34702fbf-a89c-46ed-9546-bc873caee32e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.811095 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-56b88bc7b6-8ckz6"] Dec 05 11:30:37 crc kubenswrapper[4809]: W1205 11:30:37.816076 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod232c2be5_1e83_4fc2_b2c7_9a668d572d19.slice/crio-d0170ca44a5ed898ac54a26ba30236a298a352d499632a3cc09f24a5f8a4bdf4 WatchSource:0}: Error finding container d0170ca44a5ed898ac54a26ba30236a298a352d499632a3cc09f24a5f8a4bdf4: Status 404 returned error can't find the container with id d0170ca44a5ed898ac54a26ba30236a298a352d499632a3cc09f24a5f8a4bdf4 Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.948775 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-77mnx" event={"ID":"34702fbf-a89c-46ed-9546-bc873caee32e","Type":"ContainerDied","Data":"f805ca6a121a1749dc6ffb73a6e2231313bdac33e23f2a403d0a361d785a1908"} Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.948816 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f805ca6a121a1749dc6ffb73a6e2231313bdac33e23f2a403d0a361d785a1908" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.948876 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-77mnx" Dec 05 11:30:37 crc kubenswrapper[4809]: I1205 11:30:37.951981 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b88bc7b6-8ckz6" event={"ID":"232c2be5-1e83-4fc2-b2c7-9a668d572d19","Type":"ContainerStarted","Data":"d0170ca44a5ed898ac54a26ba30236a298a352d499632a3cc09f24a5f8a4bdf4"} Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.103113 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-dbd7dc99f-5f57r"] Dec 05 11:30:38 crc kubenswrapper[4809]: E1205 11:30:38.104448 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34702fbf-a89c-46ed-9546-bc873caee32e" containerName="keystone-bootstrap" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.104470 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="34702fbf-a89c-46ed-9546-bc873caee32e" containerName="keystone-bootstrap" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.104709 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="34702fbf-a89c-46ed-9546-bc873caee32e" containerName="keystone-bootstrap" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.105514 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.109501 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.110996 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.111224 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dnjt6" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.111448 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.111622 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.111832 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.126141 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-dbd7dc99f-5f57r"] Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.209948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-scripts\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.210301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqzdf\" (UniqueName: \"kubernetes.io/projected/f96a466a-b1fb-4260-b831-e4cd814a767e-kube-api-access-vqzdf\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.210347 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-fernet-keys\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.210378 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-config-data\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.210409 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-internal-tls-certs\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.210444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-public-tls-certs\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.210477 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-combined-ca-bundle\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.210552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-credential-keys\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-public-tls-certs\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312072 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-combined-ca-bundle\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312269 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-credential-keys\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312387 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-scripts\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312420 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqzdf\" (UniqueName: \"kubernetes.io/projected/f96a466a-b1fb-4260-b831-e4cd814a767e-kube-api-access-vqzdf\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-fernet-keys\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312487 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-config-data\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.312525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-internal-tls-certs\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.318491 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-combined-ca-bundle\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.318626 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-credential-keys\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.318991 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-internal-tls-certs\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.319970 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-scripts\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.321243 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-config-data\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.321267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-public-tls-certs\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.325571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-fernet-keys\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.344596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqzdf\" (UniqueName: \"kubernetes.io/projected/f96a466a-b1fb-4260-b831-e4cd814a767e-kube-api-access-vqzdf\") pod \"keystone-dbd7dc99f-5f57r\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.495484 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.962355 4809 generic.go:334] "Generic (PLEG): container finished" podID="cc0e50e9-996a-4a27-aa6f-97bba30a0427" containerID="e66ce562a684ce2f7c9a5765f371c11805fe1afdfbe8c023db95e39baf7fd55d" exitCode=0 Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.962745 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9mn2t" event={"ID":"cc0e50e9-996a-4a27-aa6f-97bba30a0427","Type":"ContainerDied","Data":"e66ce562a684ce2f7c9a5765f371c11805fe1afdfbe8c023db95e39baf7fd55d"} Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.962828 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-dbd7dc99f-5f57r"] Dec 05 11:30:38 crc kubenswrapper[4809]: W1205 11:30:38.967969 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf96a466a_b1fb_4260_b831_e4cd814a767e.slice/crio-c4a9049e8ab1203c1ef8f173799fcb79197f9b1e43843a9524df42f27268c5f3 WatchSource:0}: Error finding container c4a9049e8ab1203c1ef8f173799fcb79197f9b1e43843a9524df42f27268c5f3: Status 404 returned error can't find the container with id c4a9049e8ab1203c1ef8f173799fcb79197f9b1e43843a9524df42f27268c5f3 Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.972572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b88bc7b6-8ckz6" event={"ID":"232c2be5-1e83-4fc2-b2c7-9a668d572d19","Type":"ContainerStarted","Data":"eb470942b72b954581fcf3a32ef1e210fa14a93a6bab02c764a013a4e22b2998"} Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.972650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b88bc7b6-8ckz6" event={"ID":"232c2be5-1e83-4fc2-b2c7-9a668d572d19","Type":"ContainerStarted","Data":"bf8cd41c788d38d0618e44a843797188ae9e6ab6ef244100c1e3de70041bbd8f"} Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.973173 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:38 crc kubenswrapper[4809]: I1205 11:30:38.974058 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.042826 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-56b88bc7b6-8ckz6" podStartSLOduration=3.042803676 podStartE2EDuration="3.042803676s" podCreationTimestamp="2025-12-05 11:30:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:39.037085546 +0000 UTC m=+1334.428062104" watchObservedRunningTime="2025-12-05 11:30:39.042803676 +0000 UTC m=+1334.433780224" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.126870 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.126918 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.202603 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.204924 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.412702 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.412795 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.457531 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.461720 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.987737 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-dbd7dc99f-5f57r" event={"ID":"f96a466a-b1fb-4260-b831-e4cd814a767e","Type":"ContainerStarted","Data":"97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d"} Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.988129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-dbd7dc99f-5f57r" event={"ID":"f96a466a-b1fb-4260-b831-e4cd814a767e","Type":"ContainerStarted","Data":"c4a9049e8ab1203c1ef8f173799fcb79197f9b1e43843a9524df42f27268c5f3"} Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.989247 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.989384 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.989529 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:39 crc kubenswrapper[4809]: I1205 11:30:39.989550 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.018768 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-dbd7dc99f-5f57r" podStartSLOduration=2.018747865 podStartE2EDuration="2.018747865s" podCreationTimestamp="2025-12-05 11:30:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:40.007933931 +0000 UTC m=+1335.398910489" watchObservedRunningTime="2025-12-05 11:30:40.018747865 +0000 UTC m=+1335.409724423" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.376540 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.456270 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-config\") pod \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.456382 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrqvl\" (UniqueName: \"kubernetes.io/projected/cc0e50e9-996a-4a27-aa6f-97bba30a0427-kube-api-access-lrqvl\") pod \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.456429 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-combined-ca-bundle\") pod \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\" (UID: \"cc0e50e9-996a-4a27-aa6f-97bba30a0427\") " Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.462535 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc0e50e9-996a-4a27-aa6f-97bba30a0427-kube-api-access-lrqvl" (OuterVolumeSpecName: "kube-api-access-lrqvl") pod "cc0e50e9-996a-4a27-aa6f-97bba30a0427" (UID: "cc0e50e9-996a-4a27-aa6f-97bba30a0427"). InnerVolumeSpecName "kube-api-access-lrqvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.487921 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-config" (OuterVolumeSpecName: "config") pod "cc0e50e9-996a-4a27-aa6f-97bba30a0427" (UID: "cc0e50e9-996a-4a27-aa6f-97bba30a0427"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.517489 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc0e50e9-996a-4a27-aa6f-97bba30a0427" (UID: "cc0e50e9-996a-4a27-aa6f-97bba30a0427"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.558761 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.558793 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrqvl\" (UniqueName: \"kubernetes.io/projected/cc0e50e9-996a-4a27-aa6f-97bba30a0427-kube-api-access-lrqvl\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.558807 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0e50e9-996a-4a27-aa6f-97bba30a0427-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.997910 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9mn2t" event={"ID":"cc0e50e9-996a-4a27-aa6f-97bba30a0427","Type":"ContainerDied","Data":"699916750a9c30999954b19d9fed3c1c85febad38b4f6917f5c1b474c4e9abc2"} Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.999007 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="699916750a9c30999954b19d9fed3c1c85febad38b4f6917f5c1b474c4e9abc2" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.999109 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:30:40 crc kubenswrapper[4809]: I1205 11:30:40.998465 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9mn2t" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.147476 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-rxkmx"] Dec 05 11:30:41 crc kubenswrapper[4809]: E1205 11:30:41.148131 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0e50e9-996a-4a27-aa6f-97bba30a0427" containerName="neutron-db-sync" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.148145 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0e50e9-996a-4a27-aa6f-97bba30a0427" containerName="neutron-db-sync" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.148330 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0e50e9-996a-4a27-aa6f-97bba30a0427" containerName="neutron-db-sync" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.149211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.173057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.173099 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-config\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.173123 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7w88\" (UniqueName: \"kubernetes.io/projected/5ec92f75-6a0d-43b1-8e4e-44be2a248295-kube-api-access-t7w88\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.173145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.173187 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.173211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.180987 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-rxkmx"] Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.227794 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-df99668dd-llrdv"] Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.236004 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.239773 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.241189 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.241344 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.242106 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-f998k" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.254921 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-df99668dd-llrdv"] Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.275823 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-config\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.275889 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.275932 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-ovndb-tls-certs\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.275953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.276010 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9klnj\" (UniqueName: \"kubernetes.io/projected/782fab1a-0354-47d4-8641-cfc7baf2bdf2-kube-api-access-9klnj\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.276039 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-combined-ca-bundle\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.276086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-httpd-config\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.276128 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.276155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-config\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.276183 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7w88\" (UniqueName: \"kubernetes.io/projected/5ec92f75-6a0d-43b1-8e4e-44be2a248295-kube-api-access-t7w88\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.276209 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.278616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.279241 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.282588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.283347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-config\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.287432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.303292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7w88\" (UniqueName: \"kubernetes.io/projected/5ec92f75-6a0d-43b1-8e4e-44be2a248295-kube-api-access-t7w88\") pod \"dnsmasq-dns-84b966f6c9-rxkmx\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.378004 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9klnj\" (UniqueName: \"kubernetes.io/projected/782fab1a-0354-47d4-8641-cfc7baf2bdf2-kube-api-access-9klnj\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.378076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-combined-ca-bundle\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.378143 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-httpd-config\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.378278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-config\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.378324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-ovndb-tls-certs\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.384127 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-httpd-config\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.384836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-ovndb-tls-certs\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.389368 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-config\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.397819 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-combined-ca-bundle\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.399295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9klnj\" (UniqueName: \"kubernetes.io/projected/782fab1a-0354-47d4-8641-cfc7baf2bdf2-kube-api-access-9klnj\") pod \"neutron-df99668dd-llrdv\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.473588 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.553403 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:41 crc kubenswrapper[4809]: I1205 11:30:41.943387 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-rxkmx"] Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.011715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" event={"ID":"5ec92f75-6a0d-43b1-8e4e-44be2a248295","Type":"ContainerStarted","Data":"51ec5e6fd05a319f6631d6c2eb43b1bb613016cd117dbc3f0fcc5309c981bcf5"} Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.587055 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-df99668dd-llrdv"] Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.798409 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.798924 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.811314 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.821676 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.821810 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:30:42 crc kubenswrapper[4809]: I1205 11:30:42.886335 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 11:30:43 crc kubenswrapper[4809]: I1205 11:30:43.064731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-df99668dd-llrdv" event={"ID":"782fab1a-0354-47d4-8641-cfc7baf2bdf2","Type":"ContainerStarted","Data":"bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4"} Dec 05 11:30:43 crc kubenswrapper[4809]: I1205 11:30:43.065126 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-df99668dd-llrdv" event={"ID":"782fab1a-0354-47d4-8641-cfc7baf2bdf2","Type":"ContainerStarted","Data":"6d53b9be2dca41fdb5b029a89d2e7eeae2a6a51c0d00c5c19efbc26b18e0cad2"} Dec 05 11:30:43 crc kubenswrapper[4809]: I1205 11:30:43.087479 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerID="068867c80f55e49b4b5a5f4b24b296d240d64a45c30b6fee625ac91d520dc067" exitCode=0 Dec 05 11:30:43 crc kubenswrapper[4809]: I1205 11:30:43.087572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" event={"ID":"5ec92f75-6a0d-43b1-8e4e-44be2a248295","Type":"ContainerDied","Data":"068867c80f55e49b4b5a5f4b24b296d240d64a45c30b6fee625ac91d520dc067"} Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.052083 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.052359 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.100024 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cffd45dc7-dpz7j"] Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.101788 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.105172 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.119109 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.119763 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cffd45dc7-dpz7j"] Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.236601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-combined-ca-bundle\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.236694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-config\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.236719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz58g\" (UniqueName: \"kubernetes.io/projected/b85b1950-6c23-424d-a724-a2a8cad9a05a-kube-api-access-lz58g\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.236742 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-public-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.236759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-ovndb-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.237014 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-httpd-config\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.237071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-internal-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.338330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-config\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.338373 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz58g\" (UniqueName: \"kubernetes.io/projected/b85b1950-6c23-424d-a724-a2a8cad9a05a-kube-api-access-lz58g\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.338398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-public-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.338413 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-ovndb-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.338510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-httpd-config\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.338537 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-internal-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.338577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-combined-ca-bundle\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.349216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-ovndb-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.349271 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-config\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.349359 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-combined-ca-bundle\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.351067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-httpd-config\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.357875 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-public-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.359225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-internal-tls-certs\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.365490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz58g\" (UniqueName: \"kubernetes.io/projected/b85b1950-6c23-424d-a724-a2a8cad9a05a-kube-api-access-lz58g\") pod \"neutron-6cffd45dc7-dpz7j\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:44 crc kubenswrapper[4809]: I1205 11:30:44.424112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:49 crc kubenswrapper[4809]: I1205 11:30:49.644793 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cffd45dc7-dpz7j"] Dec 05 11:30:49 crc kubenswrapper[4809]: W1205 11:30:49.657949 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb85b1950_6c23_424d_a724_a2a8cad9a05a.slice/crio-b714fdf18431b0dc2b38d662d0c2f76233ea078d9d327041645a4cc2e679b59a WatchSource:0}: Error finding container b714fdf18431b0dc2b38d662d0c2f76233ea078d9d327041645a4cc2e679b59a: Status 404 returned error can't find the container with id b714fdf18431b0dc2b38d662d0c2f76233ea078d9d327041645a4cc2e679b59a Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.152236 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" event={"ID":"5ec92f75-6a0d-43b1-8e4e-44be2a248295","Type":"ContainerStarted","Data":"fb2ba8a351288b4e9e91e7356e1bca3ff3d3d9546db734be03367086d01d5f39"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.152491 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.154011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2444g" event={"ID":"64f0f74c-6a4b-4549-8abb-5d497627e1bf","Type":"ContainerStarted","Data":"b355c3d21b69548d3439df18096cd1fdb75313f1444a93680f88ddda6c381b12"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.155912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffd45dc7-dpz7j" event={"ID":"b85b1950-6c23-424d-a724-a2a8cad9a05a","Type":"ContainerStarted","Data":"0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.155976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffd45dc7-dpz7j" event={"ID":"b85b1950-6c23-424d-a724-a2a8cad9a05a","Type":"ContainerStarted","Data":"987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.156060 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.156071 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffd45dc7-dpz7j" event={"ID":"b85b1950-6c23-424d-a724-a2a8cad9a05a","Type":"ContainerStarted","Data":"b714fdf18431b0dc2b38d662d0c2f76233ea078d9d327041645a4cc2e679b59a"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.157940 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-df99668dd-llrdv" event={"ID":"782fab1a-0354-47d4-8641-cfc7baf2bdf2","Type":"ContainerStarted","Data":"ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.158044 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.160703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerStarted","Data":"8c3d9899ef0e3c94f97bc4d58cbe33e5171b05c1c10d5feb8cac5ca7f8de5c14"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.160812 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.160752 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="sg-core" containerID="cri-o://be78e9bdf81a9605efc222f8084146e4de3fee13231f5195827f2c22728202ee" gracePeriod=30 Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.160748 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-central-agent" containerID="cri-o://f3a2ecf14c34773455555834aec2896c579cd69f3d3d9bbe14f6ff7e9b81d54a" gracePeriod=30 Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.160784 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-notification-agent" containerID="cri-o://4f69ac2333a4fc25a147531668dcd1cda35a5d1b56826ba9ae94f1a6d4e121da" gracePeriod=30 Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.160767 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="proxy-httpd" containerID="cri-o://8c3d9899ef0e3c94f97bc4d58cbe33e5171b05c1c10d5feb8cac5ca7f8de5c14" gracePeriod=30 Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.165275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-knlr7" event={"ID":"eb75204b-f255-48c4-bc67-54bd3601c6a9","Type":"ContainerStarted","Data":"624371613c304b126b4f3ff33d0fd6cb21766812caf78100b3abf43bd31dedf7"} Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.186425 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" podStartSLOduration=9.186400772 podStartE2EDuration="9.186400772s" podCreationTimestamp="2025-12-05 11:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:50.178867895 +0000 UTC m=+1345.569844473" watchObservedRunningTime="2025-12-05 11:30:50.186400772 +0000 UTC m=+1345.577377330" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.204953 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2444g" podStartSLOduration=2.925674514 podStartE2EDuration="44.204930908s" podCreationTimestamp="2025-12-05 11:30:06 +0000 UTC" firstStartedPulling="2025-12-05 11:30:07.67778772 +0000 UTC m=+1303.068764278" lastFinishedPulling="2025-12-05 11:30:48.957044114 +0000 UTC m=+1344.348020672" observedRunningTime="2025-12-05 11:30:50.196563729 +0000 UTC m=+1345.587540287" watchObservedRunningTime="2025-12-05 11:30:50.204930908 +0000 UTC m=+1345.595907466" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.224237 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-knlr7" podStartSLOduration=3.740435534 podStartE2EDuration="45.224215795s" podCreationTimestamp="2025-12-05 11:30:05 +0000 UTC" firstStartedPulling="2025-12-05 11:30:07.449261083 +0000 UTC m=+1302.840237641" lastFinishedPulling="2025-12-05 11:30:48.933041324 +0000 UTC m=+1344.324017902" observedRunningTime="2025-12-05 11:30:50.223768443 +0000 UTC m=+1345.614745011" watchObservedRunningTime="2025-12-05 11:30:50.224215795 +0000 UTC m=+1345.615192353" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.252475 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-df99668dd-llrdv" podStartSLOduration=9.252456516 podStartE2EDuration="9.252456516s" podCreationTimestamp="2025-12-05 11:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:50.251952022 +0000 UTC m=+1345.642928580" watchObservedRunningTime="2025-12-05 11:30:50.252456516 +0000 UTC m=+1345.643433074" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.323445 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cffd45dc7-dpz7j" podStartSLOduration=6.323424918 podStartE2EDuration="6.323424918s" podCreationTimestamp="2025-12-05 11:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:50.321377854 +0000 UTC m=+1345.712354422" watchObservedRunningTime="2025-12-05 11:30:50.323424918 +0000 UTC m=+1345.714401476" Dec 05 11:30:50 crc kubenswrapper[4809]: I1205 11:30:50.329801 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.684715492 podStartE2EDuration="45.329778105s" podCreationTimestamp="2025-12-05 11:30:05 +0000 UTC" firstStartedPulling="2025-12-05 11:30:07.436867918 +0000 UTC m=+1302.827844466" lastFinishedPulling="2025-12-05 11:30:49.081930521 +0000 UTC m=+1344.472907079" observedRunningTime="2025-12-05 11:30:50.296884071 +0000 UTC m=+1345.687860629" watchObservedRunningTime="2025-12-05 11:30:50.329778105 +0000 UTC m=+1345.720754663" Dec 05 11:30:51 crc kubenswrapper[4809]: I1205 11:30:51.177291 4809 generic.go:334] "Generic (PLEG): container finished" podID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerID="8c3d9899ef0e3c94f97bc4d58cbe33e5171b05c1c10d5feb8cac5ca7f8de5c14" exitCode=0 Dec 05 11:30:51 crc kubenswrapper[4809]: I1205 11:30:51.177591 4809 generic.go:334] "Generic (PLEG): container finished" podID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerID="be78e9bdf81a9605efc222f8084146e4de3fee13231f5195827f2c22728202ee" exitCode=2 Dec 05 11:30:51 crc kubenswrapper[4809]: I1205 11:30:51.177601 4809 generic.go:334] "Generic (PLEG): container finished" podID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerID="f3a2ecf14c34773455555834aec2896c579cd69f3d3d9bbe14f6ff7e9b81d54a" exitCode=0 Dec 05 11:30:51 crc kubenswrapper[4809]: I1205 11:30:51.177484 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerDied","Data":"8c3d9899ef0e3c94f97bc4d58cbe33e5171b05c1c10d5feb8cac5ca7f8de5c14"} Dec 05 11:30:51 crc kubenswrapper[4809]: I1205 11:30:51.177751 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerDied","Data":"be78e9bdf81a9605efc222f8084146e4de3fee13231f5195827f2c22728202ee"} Dec 05 11:30:51 crc kubenswrapper[4809]: I1205 11:30:51.177771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerDied","Data":"f3a2ecf14c34773455555834aec2896c579cd69f3d3d9bbe14f6ff7e9b81d54a"} Dec 05 11:30:52 crc kubenswrapper[4809]: I1205 11:30:52.196699 4809 generic.go:334] "Generic (PLEG): container finished" podID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerID="4f69ac2333a4fc25a147531668dcd1cda35a5d1b56826ba9ae94f1a6d4e121da" exitCode=0 Dec 05 11:30:52 crc kubenswrapper[4809]: I1205 11:30:52.196756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerDied","Data":"4f69ac2333a4fc25a147531668dcd1cda35a5d1b56826ba9ae94f1a6d4e121da"} Dec 05 11:30:52 crc kubenswrapper[4809]: I1205 11:30:52.904272 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.020672 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-log-httpd\") pod \"09271e87-1a57-43fb-81ca-8e6221dc81e0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.020984 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-run-httpd\") pod \"09271e87-1a57-43fb-81ca-8e6221dc81e0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.021159 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-sg-core-conf-yaml\") pod \"09271e87-1a57-43fb-81ca-8e6221dc81e0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.021258 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-combined-ca-bundle\") pod \"09271e87-1a57-43fb-81ca-8e6221dc81e0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.021376 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-scripts\") pod \"09271e87-1a57-43fb-81ca-8e6221dc81e0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.021431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "09271e87-1a57-43fb-81ca-8e6221dc81e0" (UID: "09271e87-1a57-43fb-81ca-8e6221dc81e0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.021471 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "09271e87-1a57-43fb-81ca-8e6221dc81e0" (UID: "09271e87-1a57-43fb-81ca-8e6221dc81e0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.021788 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-config-data\") pod \"09271e87-1a57-43fb-81ca-8e6221dc81e0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.021944 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztjkl\" (UniqueName: \"kubernetes.io/projected/09271e87-1a57-43fb-81ca-8e6221dc81e0-kube-api-access-ztjkl\") pod \"09271e87-1a57-43fb-81ca-8e6221dc81e0\" (UID: \"09271e87-1a57-43fb-81ca-8e6221dc81e0\") " Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.022651 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.022750 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09271e87-1a57-43fb-81ca-8e6221dc81e0-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.027923 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-scripts" (OuterVolumeSpecName: "scripts") pod "09271e87-1a57-43fb-81ca-8e6221dc81e0" (UID: "09271e87-1a57-43fb-81ca-8e6221dc81e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.028644 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09271e87-1a57-43fb-81ca-8e6221dc81e0-kube-api-access-ztjkl" (OuterVolumeSpecName: "kube-api-access-ztjkl") pod "09271e87-1a57-43fb-81ca-8e6221dc81e0" (UID: "09271e87-1a57-43fb-81ca-8e6221dc81e0"). InnerVolumeSpecName "kube-api-access-ztjkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.069313 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "09271e87-1a57-43fb-81ca-8e6221dc81e0" (UID: "09271e87-1a57-43fb-81ca-8e6221dc81e0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.112353 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09271e87-1a57-43fb-81ca-8e6221dc81e0" (UID: "09271e87-1a57-43fb-81ca-8e6221dc81e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.124867 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztjkl\" (UniqueName: \"kubernetes.io/projected/09271e87-1a57-43fb-81ca-8e6221dc81e0-kube-api-access-ztjkl\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.124910 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.124921 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.124936 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.153921 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-config-data" (OuterVolumeSpecName: "config-data") pod "09271e87-1a57-43fb-81ca-8e6221dc81e0" (UID: "09271e87-1a57-43fb-81ca-8e6221dc81e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.207228 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"09271e87-1a57-43fb-81ca-8e6221dc81e0","Type":"ContainerDied","Data":"feb596b2dfe233f1c3c385e2dbb04a329b8b7e7937619355edc4a270ea29f784"} Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.207298 4809 scope.go:117] "RemoveContainer" containerID="8c3d9899ef0e3c94f97bc4d58cbe33e5171b05c1c10d5feb8cac5ca7f8de5c14" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.207453 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.226242 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09271e87-1a57-43fb-81ca-8e6221dc81e0-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.237658 4809 scope.go:117] "RemoveContainer" containerID="be78e9bdf81a9605efc222f8084146e4de3fee13231f5195827f2c22728202ee" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.267823 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.273119 4809 scope.go:117] "RemoveContainer" containerID="4f69ac2333a4fc25a147531668dcd1cda35a5d1b56826ba9ae94f1a6d4e121da" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.275619 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.304395 4809 scope.go:117] "RemoveContainer" containerID="f3a2ecf14c34773455555834aec2896c579cd69f3d3d9bbe14f6ff7e9b81d54a" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.304660 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:53 crc kubenswrapper[4809]: E1205 11:30:53.310826 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="sg-core" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.310871 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="sg-core" Dec 05 11:30:53 crc kubenswrapper[4809]: E1205 11:30:53.310919 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="proxy-httpd" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.310936 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="proxy-httpd" Dec 05 11:30:53 crc kubenswrapper[4809]: E1205 11:30:53.311909 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-central-agent" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.311937 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-central-agent" Dec 05 11:30:53 crc kubenswrapper[4809]: E1205 11:30:53.311966 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-notification-agent" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.311978 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-notification-agent" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.312416 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-central-agent" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.312446 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="proxy-httpd" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.312465 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="sg-core" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.312492 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" containerName="ceilometer-notification-agent" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.316964 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.317096 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.322956 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.323157 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.429226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-log-httpd\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.429297 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-scripts\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.429482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-run-httpd\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.429592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtfxv\" (UniqueName: \"kubernetes.io/projected/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-kube-api-access-rtfxv\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.429680 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.429722 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-config-data\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.429966 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.532230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.532331 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-log-httpd\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.532393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-scripts\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.532428 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-run-httpd\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.533108 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-log-httpd\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.533284 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtfxv\" (UniqueName: \"kubernetes.io/projected/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-kube-api-access-rtfxv\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.533351 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.533384 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-config-data\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.533373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-run-httpd\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.537095 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-scripts\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.537183 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.537588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.550808 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-config-data\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.569314 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtfxv\" (UniqueName: \"kubernetes.io/projected/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-kube-api-access-rtfxv\") pod \"ceilometer-0\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " pod="openstack/ceilometer-0" Dec 05 11:30:53 crc kubenswrapper[4809]: I1205 11:30:53.670393 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:30:54 crc kubenswrapper[4809]: I1205 11:30:54.128521 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:30:54 crc kubenswrapper[4809]: I1205 11:30:54.221724 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerStarted","Data":"7ce99acbeeb400a512c03176032cc674e88522f96065b1501de93e9484899ec3"} Dec 05 11:30:54 crc kubenswrapper[4809]: I1205 11:30:54.225041 4809 generic.go:334] "Generic (PLEG): container finished" podID="64f0f74c-6a4b-4549-8abb-5d497627e1bf" containerID="b355c3d21b69548d3439df18096cd1fdb75313f1444a93680f88ddda6c381b12" exitCode=0 Dec 05 11:30:54 crc kubenswrapper[4809]: I1205 11:30:54.225069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2444g" event={"ID":"64f0f74c-6a4b-4549-8abb-5d497627e1bf","Type":"ContainerDied","Data":"b355c3d21b69548d3439df18096cd1fdb75313f1444a93680f88ddda6c381b12"} Dec 05 11:30:54 crc kubenswrapper[4809]: I1205 11:30:54.887416 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09271e87-1a57-43fb-81ca-8e6221dc81e0" path="/var/lib/kubelet/pods/09271e87-1a57-43fb-81ca-8e6221dc81e0/volumes" Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.235224 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerStarted","Data":"3eceec205c19750415c3ff9e7f89e9aeb54d239bff3973eb107f4253f2992346"} Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.587980 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.674228 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8b7s\" (UniqueName: \"kubernetes.io/projected/64f0f74c-6a4b-4549-8abb-5d497627e1bf-kube-api-access-t8b7s\") pod \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.675389 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-combined-ca-bundle\") pod \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.675938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-db-sync-config-data\") pod \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\" (UID: \"64f0f74c-6a4b-4549-8abb-5d497627e1bf\") " Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.680607 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "64f0f74c-6a4b-4549-8abb-5d497627e1bf" (UID: "64f0f74c-6a4b-4549-8abb-5d497627e1bf"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.681053 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64f0f74c-6a4b-4549-8abb-5d497627e1bf-kube-api-access-t8b7s" (OuterVolumeSpecName: "kube-api-access-t8b7s") pod "64f0f74c-6a4b-4549-8abb-5d497627e1bf" (UID: "64f0f74c-6a4b-4549-8abb-5d497627e1bf"). InnerVolumeSpecName "kube-api-access-t8b7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.704997 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64f0f74c-6a4b-4549-8abb-5d497627e1bf" (UID: "64f0f74c-6a4b-4549-8abb-5d497627e1bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.779458 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8b7s\" (UniqueName: \"kubernetes.io/projected/64f0f74c-6a4b-4549-8abb-5d497627e1bf-kube-api-access-t8b7s\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.779501 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:55 crc kubenswrapper[4809]: I1205 11:30:55.779515 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/64f0f74c-6a4b-4549-8abb-5d497627e1bf-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.246759 4809 generic.go:334] "Generic (PLEG): container finished" podID="eb75204b-f255-48c4-bc67-54bd3601c6a9" containerID="624371613c304b126b4f3ff33d0fd6cb21766812caf78100b3abf43bd31dedf7" exitCode=0 Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.246932 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-knlr7" event={"ID":"eb75204b-f255-48c4-bc67-54bd3601c6a9","Type":"ContainerDied","Data":"624371613c304b126b4f3ff33d0fd6cb21766812caf78100b3abf43bd31dedf7"} Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.251240 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2444g" event={"ID":"64f0f74c-6a4b-4549-8abb-5d497627e1bf","Type":"ContainerDied","Data":"a33323980bb5de72b98b3a0ed378804abfac195584dd0685e9c61d551d3aa1e8"} Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.251273 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a33323980bb5de72b98b3a0ed378804abfac195584dd0685e9c61d551d3aa1e8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.251278 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2444g" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.256464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerStarted","Data":"e24f45701e293ee5b251a98a42d425ad6a6a5c84c03dc03ad9d44e980855289d"} Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.474882 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.521669 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68b55f4bc-9ngw8"] Dec 05 11:30:56 crc kubenswrapper[4809]: E1205 11:30:56.525482 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64f0f74c-6a4b-4549-8abb-5d497627e1bf" containerName="barbican-db-sync" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.525575 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="64f0f74c-6a4b-4549-8abb-5d497627e1bf" containerName="barbican-db-sync" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.525823 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="64f0f74c-6a4b-4549-8abb-5d497627e1bf" containerName="barbican-db-sync" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.526749 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.535424 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.536815 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-n5dxb" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.547190 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b55f4bc-9ngw8"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.550149 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.576528 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-69c49cc5f8-w57p4"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.578163 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.597226 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.599184 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.599254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data-custom\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.599346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-combined-ca-bundle\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.599387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ttt9\" (UniqueName: \"kubernetes.io/projected/95c323f7-b5da-4767-96dd-1c96f82c774d-kube-api-access-4ttt9\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.599468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c323f7-b5da-4767-96dd-1c96f82c774d-logs\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.663026 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-69c49cc5f8-w57p4"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.681484 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-4hr9q"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.681816 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" podUID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerName="dnsmasq-dns" containerID="cri-o://9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521" gracePeriod=10 Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data-custom\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700732 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-combined-ca-bundle\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ttt9\" (UniqueName: \"kubernetes.io/projected/95c323f7-b5da-4767-96dd-1c96f82c774d-kube-api-access-4ttt9\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data-custom\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700820 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf66w\" (UniqueName: \"kubernetes.io/projected/919e76fa-7915-4a0f-8482-b827879dcacb-kube-api-access-bf66w\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700842 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c323f7-b5da-4767-96dd-1c96f82c774d-logs\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700894 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-combined-ca-bundle\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700936 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.700954 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919e76fa-7915-4a0f-8482-b827879dcacb-logs\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.701407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c323f7-b5da-4767-96dd-1c96f82c774d-logs\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.707922 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.713516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data-custom\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.730303 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ttt9\" (UniqueName: \"kubernetes.io/projected/95c323f7-b5da-4767-96dd-1c96f82c774d-kube-api-access-4ttt9\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.770744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-combined-ca-bundle\") pod \"barbican-worker-68b55f4bc-9ngw8\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.802955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data-custom\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.803016 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf66w\" (UniqueName: \"kubernetes.io/projected/919e76fa-7915-4a0f-8482-b827879dcacb-kube-api-access-bf66w\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.803039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.803170 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-combined-ca-bundle\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.803230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919e76fa-7915-4a0f-8482-b827879dcacb-logs\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.803647 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919e76fa-7915-4a0f-8482-b827879dcacb-logs\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.805367 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2l6kn"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.807109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.807126 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data-custom\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.827436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-combined-ca-bundle\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.827514 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7cb9769bdb-8tnj5"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.829126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.830559 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.832938 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf66w\" (UniqueName: \"kubernetes.io/projected/919e76fa-7915-4a0f-8482-b827879dcacb-kube-api-access-bf66w\") pod \"barbican-keystone-listener-69c49cc5f8-w57p4\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.836448 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.843462 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2l6kn"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.850694 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7cb9769bdb-8tnj5"] Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.856535 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.905510 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.905557 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwr72\" (UniqueName: \"kubernetes.io/projected/80cac454-9282-41a3-b376-c2beb5b0096c-kube-api-access-pwr72\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.905736 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.905837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80cac454-9282-41a3-b376-c2beb5b0096c-logs\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.905937 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.906065 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.906174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.906243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-combined-ca-bundle\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.906261 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-config\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.906322 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fvtb\" (UniqueName: \"kubernetes.io/projected/f460a24a-b9da-488f-8379-ae952cdee06e-kube-api-access-7fvtb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.906351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data-custom\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:56 crc kubenswrapper[4809]: I1205 11:30:56.930199 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.009795 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010244 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-combined-ca-bundle\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010267 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-config\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010313 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fvtb\" (UniqueName: \"kubernetes.io/projected/f460a24a-b9da-488f-8379-ae952cdee06e-kube-api-access-7fvtb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data-custom\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010398 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwr72\" (UniqueName: \"kubernetes.io/projected/80cac454-9282-41a3-b376-c2beb5b0096c-kube-api-access-pwr72\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010463 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010494 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80cac454-9282-41a3-b376-c2beb5b0096c-logs\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010534 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.010986 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.011270 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-config\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.011399 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.011587 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80cac454-9282-41a3-b376-c2beb5b0096c-logs\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.012026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.014561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.015585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data-custom\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.017104 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-combined-ca-bundle\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.018169 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.031677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwr72\" (UniqueName: \"kubernetes.io/projected/80cac454-9282-41a3-b376-c2beb5b0096c-kube-api-access-pwr72\") pod \"barbican-api-7cb9769bdb-8tnj5\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.034528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fvtb\" (UniqueName: \"kubernetes.io/projected/f460a24a-b9da-488f-8379-ae952cdee06e-kube-api-access-7fvtb\") pod \"dnsmasq-dns-75c8ddd69c-2l6kn\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.224258 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.252371 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.279682 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.306188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerStarted","Data":"a7544ff88a950bd54ea465532a74b580658be779670e8973cb5b65032dcf574f"} Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.309259 4809 generic.go:334] "Generic (PLEG): container finished" podID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerID="9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521" exitCode=0 Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.309453 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.309734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" event={"ID":"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9","Type":"ContainerDied","Data":"9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521"} Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.309811 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-4hr9q" event={"ID":"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9","Type":"ContainerDied","Data":"b641879e2a9eea87d98a95f5f5a9278843966d4ad136f2d147572070e80e2c74"} Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.309849 4809 scope.go:117] "RemoveContainer" containerID="9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.316815 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl2v9\" (UniqueName: \"kubernetes.io/projected/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-kube-api-access-fl2v9\") pod \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.316875 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-swift-storage-0\") pod \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.316940 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-config\") pod \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.316995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-svc\") pod \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.317022 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-sb\") pod \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.317048 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-nb\") pod \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\" (UID: \"9e3fe5e5-aeaa-402c-8b90-715656ecf6d9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.367710 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-kube-api-access-fl2v9" (OuterVolumeSpecName: "kube-api-access-fl2v9") pod "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" (UID: "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9"). InnerVolumeSpecName "kube-api-access-fl2v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.418993 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl2v9\" (UniqueName: \"kubernetes.io/projected/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-kube-api-access-fl2v9\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.448196 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b55f4bc-9ngw8"] Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.473486 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-69c49cc5f8-w57p4"] Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.495355 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" (UID: "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.503808 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" (UID: "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.507356 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-config" (OuterVolumeSpecName: "config") pod "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" (UID: "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.515213 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" (UID: "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.520384 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.520419 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.520429 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.520445 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.530697 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" (UID: "9e3fe5e5-aeaa-402c-8b90-715656ecf6d9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.615036 4809 scope.go:117] "RemoveContainer" containerID="d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.622423 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.676970 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-4hr9q"] Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.716099 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-4hr9q"] Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.745170 4809 scope.go:117] "RemoveContainer" containerID="9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521" Dec 05 11:30:57 crc kubenswrapper[4809]: E1205 11:30:57.745513 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521\": container with ID starting with 9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521 not found: ID does not exist" containerID="9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.745555 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521"} err="failed to get container status \"9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521\": rpc error: code = NotFound desc = could not find container \"9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521\": container with ID starting with 9041ae0e067e2ec90d3a001b09ae2c0eb5696d8f93d8b98967c327ca2ee4e521 not found: ID does not exist" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.745581 4809 scope.go:117] "RemoveContainer" containerID="d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa" Dec 05 11:30:57 crc kubenswrapper[4809]: E1205 11:30:57.746141 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa\": container with ID starting with d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa not found: ID does not exist" containerID="d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.746182 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa"} err="failed to get container status \"d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa\": rpc error: code = NotFound desc = could not find container \"d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa\": container with ID starting with d5a8437454d1e7b3804dac38cb094a10b38cda058b528459d54a5b2d2eb534aa not found: ID does not exist" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.786346 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.825328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb75204b-f255-48c4-bc67-54bd3601c6a9-etc-machine-id\") pod \"eb75204b-f255-48c4-bc67-54bd3601c6a9\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.825570 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-config-data\") pod \"eb75204b-f255-48c4-bc67-54bd3601c6a9\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.825455 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb75204b-f255-48c4-bc67-54bd3601c6a9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "eb75204b-f255-48c4-bc67-54bd3601c6a9" (UID: "eb75204b-f255-48c4-bc67-54bd3601c6a9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.825917 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-combined-ca-bundle\") pod \"eb75204b-f255-48c4-bc67-54bd3601c6a9\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.826049 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-db-sync-config-data\") pod \"eb75204b-f255-48c4-bc67-54bd3601c6a9\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.826321 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-scripts\") pod \"eb75204b-f255-48c4-bc67-54bd3601c6a9\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.826440 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp9tm\" (UniqueName: \"kubernetes.io/projected/eb75204b-f255-48c4-bc67-54bd3601c6a9-kube-api-access-zp9tm\") pod \"eb75204b-f255-48c4-bc67-54bd3601c6a9\" (UID: \"eb75204b-f255-48c4-bc67-54bd3601c6a9\") " Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.826926 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb75204b-f255-48c4-bc67-54bd3601c6a9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.831122 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eb75204b-f255-48c4-bc67-54bd3601c6a9" (UID: "eb75204b-f255-48c4-bc67-54bd3601c6a9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.837823 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-scripts" (OuterVolumeSpecName: "scripts") pod "eb75204b-f255-48c4-bc67-54bd3601c6a9" (UID: "eb75204b-f255-48c4-bc67-54bd3601c6a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.853866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb75204b-f255-48c4-bc67-54bd3601c6a9-kube-api-access-zp9tm" (OuterVolumeSpecName: "kube-api-access-zp9tm") pod "eb75204b-f255-48c4-bc67-54bd3601c6a9" (UID: "eb75204b-f255-48c4-bc67-54bd3601c6a9"). InnerVolumeSpecName "kube-api-access-zp9tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.857890 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb75204b-f255-48c4-bc67-54bd3601c6a9" (UID: "eb75204b-f255-48c4-bc67-54bd3601c6a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.888742 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-config-data" (OuterVolumeSpecName: "config-data") pod "eb75204b-f255-48c4-bc67-54bd3601c6a9" (UID: "eb75204b-f255-48c4-bc67-54bd3601c6a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.929151 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.929191 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.929203 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.929216 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp9tm\" (UniqueName: \"kubernetes.io/projected/eb75204b-f255-48c4-bc67-54bd3601c6a9-kube-api-access-zp9tm\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:57 crc kubenswrapper[4809]: I1205 11:30:57.929229 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb75204b-f255-48c4-bc67-54bd3601c6a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.017305 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7cb9769bdb-8tnj5"] Dec 05 11:30:58 crc kubenswrapper[4809]: W1205 11:30:58.020973 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf460a24a_b9da_488f_8379_ae952cdee06e.slice/crio-d0c9f0fc09cd77fcc741204b349cf9d1d275db0c871c7441964b346884cbabf3 WatchSource:0}: Error finding container d0c9f0fc09cd77fcc741204b349cf9d1d275db0c871c7441964b346884cbabf3: Status 404 returned error can't find the container with id d0c9f0fc09cd77fcc741204b349cf9d1d275db0c871c7441964b346884cbabf3 Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.043182 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2l6kn"] Dec 05 11:30:58 crc kubenswrapper[4809]: W1205 11:30:58.047261 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80cac454_9282_41a3_b376_c2beb5b0096c.slice/crio-c68105d6e46325de7ac1d4835c42b9b2d894c8cd41bf640e01934c67c5ba2b78 WatchSource:0}: Error finding container c68105d6e46325de7ac1d4835c42b9b2d894c8cd41bf640e01934c67c5ba2b78: Status 404 returned error can't find the container with id c68105d6e46325de7ac1d4835c42b9b2d894c8cd41bf640e01934c67c5ba2b78 Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.326762 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" event={"ID":"f460a24a-b9da-488f-8379-ae952cdee06e","Type":"ContainerStarted","Data":"2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.327123 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" event={"ID":"f460a24a-b9da-488f-8379-ae952cdee06e","Type":"ContainerStarted","Data":"d0c9f0fc09cd77fcc741204b349cf9d1d275db0c871c7441964b346884cbabf3"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.332712 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" event={"ID":"919e76fa-7915-4a0f-8482-b827879dcacb","Type":"ContainerStarted","Data":"96e87795672f5b2e4cfc92bc314b4cd446ceb9e5767b729b4440e87f4c902eeb"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.360476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b55f4bc-9ngw8" event={"ID":"95c323f7-b5da-4767-96dd-1c96f82c774d","Type":"ContainerStarted","Data":"bc458e54ed56104cf41e973c977e2857cf8012cce3d3655c23cf3146f7117faa"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.377442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cb9769bdb-8tnj5" event={"ID":"80cac454-9282-41a3-b376-c2beb5b0096c","Type":"ContainerStarted","Data":"aa267d76d96ed36f44971b8d32186525836ead41231817f2b83d510110f590cb"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.377509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cb9769bdb-8tnj5" event={"ID":"80cac454-9282-41a3-b376-c2beb5b0096c","Type":"ContainerStarted","Data":"c68105d6e46325de7ac1d4835c42b9b2d894c8cd41bf640e01934c67c5ba2b78"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.391820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerStarted","Data":"f57b5e787af7f2f1fb584535e6c2918eab3d4b7aed79704f311d5c22e1ee9e22"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.392215 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.394219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-knlr7" event={"ID":"eb75204b-f255-48c4-bc67-54bd3601c6a9","Type":"ContainerDied","Data":"aecfb9c6ed9754ed7fc6d8cdda5130e3addb4d34057aba9f979823e4f291c6de"} Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.394272 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aecfb9c6ed9754ed7fc6d8cdda5130e3addb4d34057aba9f979823e4f291c6de" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.394285 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-knlr7" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.428464 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.806976948 podStartE2EDuration="5.42843817s" podCreationTimestamp="2025-12-05 11:30:53 +0000 UTC" firstStartedPulling="2025-12-05 11:30:54.14660353 +0000 UTC m=+1349.537580088" lastFinishedPulling="2025-12-05 11:30:57.768064752 +0000 UTC m=+1353.159041310" observedRunningTime="2025-12-05 11:30:58.421422336 +0000 UTC m=+1353.812398904" watchObservedRunningTime="2025-12-05 11:30:58.42843817 +0000 UTC m=+1353.819414728" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.549887 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:30:58 crc kubenswrapper[4809]: E1205 11:30:58.550230 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerName="dnsmasq-dns" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.550241 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerName="dnsmasq-dns" Dec 05 11:30:58 crc kubenswrapper[4809]: E1205 11:30:58.550255 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerName="init" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.550261 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerName="init" Dec 05 11:30:58 crc kubenswrapper[4809]: E1205 11:30:58.550279 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb75204b-f255-48c4-bc67-54bd3601c6a9" containerName="cinder-db-sync" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.550285 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb75204b-f255-48c4-bc67-54bd3601c6a9" containerName="cinder-db-sync" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.550442 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb75204b-f255-48c4-bc67-54bd3601c6a9" containerName="cinder-db-sync" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.550457 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" containerName="dnsmasq-dns" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.551362 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.557469 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.558479 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.558660 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q2l2q" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.559520 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.569124 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.612905 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2l6kn"] Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.646608 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.646693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2557186-1276-41f5-99af-571e2db4c44e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.646717 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.646731 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pngsm\" (UniqueName: \"kubernetes.io/projected/a2557186-1276-41f5-99af-571e2db4c44e-kube-api-access-pngsm\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.646785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-scripts\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.646824 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.728925 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-8h6tp"] Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.738726 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749282 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2557186-1276-41f5-99af-571e2db4c44e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749321 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pngsm\" (UniqueName: \"kubernetes.io/projected/a2557186-1276-41f5-99af-571e2db4c44e-kube-api-access-pngsm\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-scripts\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-config\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2557186-1276-41f5-99af-571e2db4c44e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749464 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749560 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.749882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.750708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq6fh\" (UniqueName: \"kubernetes.io/projected/d52c2032-df32-4ff3-ab67-3c623e6fe80c-kube-api-access-jq6fh\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.750731 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.750846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.799457 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-8h6tp"] Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.895691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.895762 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.895833 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq6fh\" (UniqueName: \"kubernetes.io/projected/d52c2032-df32-4ff3-ab67-3c623e6fe80c-kube-api-access-jq6fh\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.895854 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.895975 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.896038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-config\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.896948 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-config\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.897570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.898162 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.900472 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.901260 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.924973 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e3fe5e5-aeaa-402c-8b90-715656ecf6d9" path="/var/lib/kubelet/pods/9e3fe5e5-aeaa-402c-8b90-715656ecf6d9/volumes" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.925706 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.931183 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.937943 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.940103 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.958927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.965528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.969330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-scripts\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.972323 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.978371 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq6fh\" (UniqueName: \"kubernetes.io/projected/d52c2032-df32-4ff3-ab67-3c623e6fe80c-kube-api-access-jq6fh\") pod \"dnsmasq-dns-5784cf869f-8h6tp\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.978373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pngsm\" (UniqueName: \"kubernetes.io/projected/a2557186-1276-41f5-99af-571e2db4c44e-kube-api-access-pngsm\") pod \"cinder-scheduler-0\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " pod="openstack/cinder-scheduler-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.999938 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:58 crc kubenswrapper[4809]: I1205 11:30:58.999991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7xtq\" (UniqueName: \"kubernetes.io/projected/9aa5170f-d144-4731-adba-c25a63f36ac7-kube-api-access-f7xtq\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.000024 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data-custom\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.000066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9aa5170f-d144-4731-adba-c25a63f36ac7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.000091 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-scripts\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.000139 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.000190 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa5170f-d144-4731-adba-c25a63f36ac7-logs\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.089115 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.102760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.102846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7xtq\" (UniqueName: \"kubernetes.io/projected/9aa5170f-d144-4731-adba-c25a63f36ac7-kube-api-access-f7xtq\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.102882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data-custom\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.102933 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9aa5170f-d144-4731-adba-c25a63f36ac7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.102956 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-scripts\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.103002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.103050 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa5170f-d144-4731-adba-c25a63f36ac7-logs\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.103546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa5170f-d144-4731-adba-c25a63f36ac7-logs\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.103590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9aa5170f-d144-4731-adba-c25a63f36ac7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.107659 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-scripts\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.108215 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.109344 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.109974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data-custom\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.142006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7xtq\" (UniqueName: \"kubernetes.io/projected/9aa5170f-d144-4731-adba-c25a63f36ac7-kube-api-access-f7xtq\") pod \"cinder-api-0\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.164949 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.181550 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.407373 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cb9769bdb-8tnj5" event={"ID":"80cac454-9282-41a3-b376-c2beb5b0096c","Type":"ContainerStarted","Data":"a5c195143bfe59cebbb4f4a0366446d0f7ba8c9aa9a37e99673ab8da024eba8f"} Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.407661 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.416664 4809 generic.go:334] "Generic (PLEG): container finished" podID="f460a24a-b9da-488f-8379-ae952cdee06e" containerID="2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4" exitCode=0 Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.416720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" event={"ID":"f460a24a-b9da-488f-8379-ae952cdee06e","Type":"ContainerDied","Data":"2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4"} Dec 05 11:30:59 crc kubenswrapper[4809]: I1205 11:30:59.433096 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7cb9769bdb-8tnj5" podStartSLOduration=3.4330781630000002 podStartE2EDuration="3.433078163s" podCreationTimestamp="2025-12-05 11:30:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:30:59.425184616 +0000 UTC m=+1354.816161174" watchObservedRunningTime="2025-12-05 11:30:59.433078163 +0000 UTC m=+1354.824054721" Dec 05 11:31:00 crc kubenswrapper[4809]: I1205 11:31:00.434828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" event={"ID":"f460a24a-b9da-488f-8379-ae952cdee06e","Type":"ContainerStarted","Data":"80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862"} Dec 05 11:31:00 crc kubenswrapper[4809]: I1205 11:31:00.439046 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b55f4bc-9ngw8" event={"ID":"95c323f7-b5da-4767-96dd-1c96f82c774d","Type":"ContainerStarted","Data":"ffa1c61cecbaf03fa197f25a2a9928b22008f344ca29a0aacbc7194bddad9bc7"} Dec 05 11:31:00 crc kubenswrapper[4809]: I1205 11:31:00.439201 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:31:00 crc kubenswrapper[4809]: I1205 11:31:00.451774 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:31:00 crc kubenswrapper[4809]: W1205 11:31:00.464628 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aa5170f_d144_4731_adba_c25a63f36ac7.slice/crio-c5863d3c54b16a3daf7a216ae8123ecd8cf7328974446cb6a158add817d88bf4 WatchSource:0}: Error finding container c5863d3c54b16a3daf7a216ae8123ecd8cf7328974446cb6a158add817d88bf4: Status 404 returned error can't find the container with id c5863d3c54b16a3daf7a216ae8123ecd8cf7328974446cb6a158add817d88bf4 Dec 05 11:31:00 crc kubenswrapper[4809]: I1205 11:31:00.660745 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-8h6tp"] Dec 05 11:31:00 crc kubenswrapper[4809]: I1205 11:31:00.696384 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:31:00 crc kubenswrapper[4809]: W1205 11:31:00.699730 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd52c2032_df32_4ff3_ab67_3c623e6fe80c.slice/crio-141e7c8d0fe6b843b8eef6a5be1fdfccb90c019b85d296aa18caf1f74fa6f2bb WatchSource:0}: Error finding container 141e7c8d0fe6b843b8eef6a5be1fdfccb90c019b85d296aa18caf1f74fa6f2bb: Status 404 returned error can't find the container with id 141e7c8d0fe6b843b8eef6a5be1fdfccb90c019b85d296aa18caf1f74fa6f2bb Dec 05 11:31:00 crc kubenswrapper[4809]: W1205 11:31:00.706085 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2557186_1276_41f5_99af_571e2db4c44e.slice/crio-2ff6fc468d4b53b1e09db2073a4390bf7b073f5d675cd002d5ee8ec42cfebe84 WatchSource:0}: Error finding container 2ff6fc468d4b53b1e09db2073a4390bf7b073f5d675cd002d5ee8ec42cfebe84: Status 404 returned error can't find the container with id 2ff6fc468d4b53b1e09db2073a4390bf7b073f5d675cd002d5ee8ec42cfebe84 Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.470781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a2557186-1276-41f5-99af-571e2db4c44e","Type":"ContainerStarted","Data":"2ff6fc468d4b53b1e09db2073a4390bf7b073f5d675cd002d5ee8ec42cfebe84"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.477511 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b55f4bc-9ngw8" event={"ID":"95c323f7-b5da-4767-96dd-1c96f82c774d","Type":"ContainerStarted","Data":"929972b6f8b7d1846e06117ff7d6fe1f452d8c49bdedcc3dadd82bba1e6d2dc7"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.481351 4809 generic.go:334] "Generic (PLEG): container finished" podID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerID="88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32" exitCode=0 Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.481503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" event={"ID":"d52c2032-df32-4ff3-ab67-3c623e6fe80c","Type":"ContainerDied","Data":"88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.481554 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" event={"ID":"d52c2032-df32-4ff3-ab67-3c623e6fe80c","Type":"ContainerStarted","Data":"141e7c8d0fe6b843b8eef6a5be1fdfccb90c019b85d296aa18caf1f74fa6f2bb"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.487617 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9aa5170f-d144-4731-adba-c25a63f36ac7","Type":"ContainerStarted","Data":"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.487686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9aa5170f-d144-4731-adba-c25a63f36ac7","Type":"ContainerStarted","Data":"c5863d3c54b16a3daf7a216ae8123ecd8cf7328974446cb6a158add817d88bf4"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.492930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" event={"ID":"919e76fa-7915-4a0f-8482-b827879dcacb","Type":"ContainerStarted","Data":"bd369f5a80a27b3f543389e7e7b300b6bde366ca4736f6ed6dc27ce406afa252"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.492971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" event={"ID":"919e76fa-7915-4a0f-8482-b827879dcacb","Type":"ContainerStarted","Data":"c9e36f881b5958ccc505a0d69b7688ee45bfcd1841d3987cabc75687e7c369c0"} Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.493204 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" podUID="f460a24a-b9da-488f-8379-ae952cdee06e" containerName="dnsmasq-dns" containerID="cri-o://80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862" gracePeriod=10 Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.493320 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.549776 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68b55f4bc-9ngw8" podStartSLOduration=2.9583120149999997 podStartE2EDuration="5.549608193s" podCreationTimestamp="2025-12-05 11:30:56 +0000 UTC" firstStartedPulling="2025-12-05 11:30:57.485729073 +0000 UTC m=+1352.876705641" lastFinishedPulling="2025-12-05 11:31:00.077025261 +0000 UTC m=+1355.468001819" observedRunningTime="2025-12-05 11:31:01.503307468 +0000 UTC m=+1356.894284036" watchObservedRunningTime="2025-12-05 11:31:01.549608193 +0000 UTC m=+1356.940584751" Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.585972 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" podStartSLOduration=5.5859505160000005 podStartE2EDuration="5.585950516s" podCreationTimestamp="2025-12-05 11:30:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:01.569191716 +0000 UTC m=+1356.960168274" watchObservedRunningTime="2025-12-05 11:31:01.585950516 +0000 UTC m=+1356.976927074" Dec 05 11:31:01 crc kubenswrapper[4809]: I1205 11:31:01.601891 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" podStartSLOduration=2.974044997 podStartE2EDuration="5.601855254s" podCreationTimestamp="2025-12-05 11:30:56 +0000 UTC" firstStartedPulling="2025-12-05 11:30:57.502359479 +0000 UTC m=+1352.893336037" lastFinishedPulling="2025-12-05 11:31:00.130169736 +0000 UTC m=+1355.521146294" observedRunningTime="2025-12-05 11:31:01.597875469 +0000 UTC m=+1356.988852047" watchObservedRunningTime="2025-12-05 11:31:01.601855254 +0000 UTC m=+1356.992831812" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.050202 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.173011 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-swift-storage-0\") pod \"f460a24a-b9da-488f-8379-ae952cdee06e\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.173268 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fvtb\" (UniqueName: \"kubernetes.io/projected/f460a24a-b9da-488f-8379-ae952cdee06e-kube-api-access-7fvtb\") pod \"f460a24a-b9da-488f-8379-ae952cdee06e\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.173326 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-sb\") pod \"f460a24a-b9da-488f-8379-ae952cdee06e\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.173383 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-nb\") pod \"f460a24a-b9da-488f-8379-ae952cdee06e\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.173419 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-config\") pod \"f460a24a-b9da-488f-8379-ae952cdee06e\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.173504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-svc\") pod \"f460a24a-b9da-488f-8379-ae952cdee06e\" (UID: \"f460a24a-b9da-488f-8379-ae952cdee06e\") " Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.189017 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f460a24a-b9da-488f-8379-ae952cdee06e-kube-api-access-7fvtb" (OuterVolumeSpecName: "kube-api-access-7fvtb") pod "f460a24a-b9da-488f-8379-ae952cdee06e" (UID: "f460a24a-b9da-488f-8379-ae952cdee06e"). InnerVolumeSpecName "kube-api-access-7fvtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.230665 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f460a24a-b9da-488f-8379-ae952cdee06e" (UID: "f460a24a-b9da-488f-8379-ae952cdee06e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.237150 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f460a24a-b9da-488f-8379-ae952cdee06e" (UID: "f460a24a-b9da-488f-8379-ae952cdee06e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.248021 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f460a24a-b9da-488f-8379-ae952cdee06e" (UID: "f460a24a-b9da-488f-8379-ae952cdee06e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.250822 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f460a24a-b9da-488f-8379-ae952cdee06e" (UID: "f460a24a-b9da-488f-8379-ae952cdee06e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.265331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-config" (OuterVolumeSpecName: "config") pod "f460a24a-b9da-488f-8379-ae952cdee06e" (UID: "f460a24a-b9da-488f-8379-ae952cdee06e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.277686 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.277717 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.277727 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fvtb\" (UniqueName: \"kubernetes.io/projected/f460a24a-b9da-488f-8379-ae952cdee06e-kube-api-access-7fvtb\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.277736 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.277744 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.277752 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f460a24a-b9da-488f-8379-ae952cdee06e-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.531401 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" event={"ID":"d52c2032-df32-4ff3-ab67-3c623e6fe80c","Type":"ContainerStarted","Data":"41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4"} Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.531732 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.535035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9aa5170f-d144-4731-adba-c25a63f36ac7","Type":"ContainerStarted","Data":"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad"} Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.535184 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.544402 4809 generic.go:334] "Generic (PLEG): container finished" podID="f460a24a-b9da-488f-8379-ae952cdee06e" containerID="80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862" exitCode=0 Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.544464 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.544452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" event={"ID":"f460a24a-b9da-488f-8379-ae952cdee06e","Type":"ContainerDied","Data":"80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862"} Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.544599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-2l6kn" event={"ID":"f460a24a-b9da-488f-8379-ae952cdee06e","Type":"ContainerDied","Data":"d0c9f0fc09cd77fcc741204b349cf9d1d275db0c871c7441964b346884cbabf3"} Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.544687 4809 scope.go:117] "RemoveContainer" containerID="80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.548561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a2557186-1276-41f5-99af-571e2db4c44e","Type":"ContainerStarted","Data":"f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5"} Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.554469 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" podStartSLOduration=4.554448611 podStartE2EDuration="4.554448611s" podCreationTimestamp="2025-12-05 11:30:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:02.55328954 +0000 UTC m=+1357.944266108" watchObservedRunningTime="2025-12-05 11:31:02.554448611 +0000 UTC m=+1357.945425169" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.580402 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.580383031 podStartE2EDuration="4.580383031s" podCreationTimestamp="2025-12-05 11:30:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:02.567993216 +0000 UTC m=+1357.958969794" watchObservedRunningTime="2025-12-05 11:31:02.580383031 +0000 UTC m=+1357.971359589" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.597945 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2l6kn"] Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.603582 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-2l6kn"] Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.614763 4809 scope.go:117] "RemoveContainer" containerID="2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.638185 4809 scope.go:117] "RemoveContainer" containerID="80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862" Dec 05 11:31:02 crc kubenswrapper[4809]: E1205 11:31:02.639141 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862\": container with ID starting with 80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862 not found: ID does not exist" containerID="80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.639199 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862"} err="failed to get container status \"80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862\": rpc error: code = NotFound desc = could not find container \"80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862\": container with ID starting with 80c14b591c6a9165ce9618e7ee6a56cd10bdf0804e31befc25ba03bc86016862 not found: ID does not exist" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.639233 4809 scope.go:117] "RemoveContainer" containerID="2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4" Dec 05 11:31:02 crc kubenswrapper[4809]: E1205 11:31:02.639774 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4\": container with ID starting with 2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4 not found: ID does not exist" containerID="2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.639867 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4"} err="failed to get container status \"2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4\": rpc error: code = NotFound desc = could not find container \"2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4\": container with ID starting with 2da216bec0020a189da94833e417e8ef8ad2954fa949a900664fb06f7cc3f4f4 not found: ID does not exist" Dec 05 11:31:02 crc kubenswrapper[4809]: I1205 11:31:02.887563 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f460a24a-b9da-488f-8379-ae952cdee06e" path="/var/lib/kubelet/pods/f460a24a-b9da-488f-8379-ae952cdee06e/volumes" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.431104 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.559852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a2557186-1276-41f5-99af-571e2db4c44e","Type":"ContainerStarted","Data":"bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d"} Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.584673 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.855675535 podStartE2EDuration="5.584650384s" podCreationTimestamp="2025-12-05 11:30:58 +0000 UTC" firstStartedPulling="2025-12-05 11:31:00.707551556 +0000 UTC m=+1356.098528114" lastFinishedPulling="2025-12-05 11:31:01.436526405 +0000 UTC m=+1356.827502963" observedRunningTime="2025-12-05 11:31:03.577483576 +0000 UTC m=+1358.968460144" watchObservedRunningTime="2025-12-05 11:31:03.584650384 +0000 UTC m=+1358.975626972" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.961779 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-847649ff46-ldvld"] Dec 05 11:31:03 crc kubenswrapper[4809]: E1205 11:31:03.962507 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f460a24a-b9da-488f-8379-ae952cdee06e" containerName="dnsmasq-dns" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.962616 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f460a24a-b9da-488f-8379-ae952cdee06e" containerName="dnsmasq-dns" Dec 05 11:31:03 crc kubenswrapper[4809]: E1205 11:31:03.962744 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f460a24a-b9da-488f-8379-ae952cdee06e" containerName="init" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.962822 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f460a24a-b9da-488f-8379-ae952cdee06e" containerName="init" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.963118 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f460a24a-b9da-488f-8379-ae952cdee06e" containerName="dnsmasq-dns" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.964144 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.971342 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.971567 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 05 11:31:03 crc kubenswrapper[4809]: I1205 11:31:03.982475 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-847649ff46-ldvld"] Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.007290 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-public-tls-certs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.007332 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data-custom\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.007349 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-internal-tls-certs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.007380 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.007406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755354bd-175b-46e6-a9bf-93d8b9eab885-logs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.007420 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-combined-ca-bundle\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.007480 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq692\" (UniqueName: \"kubernetes.io/projected/755354bd-175b-46e6-a9bf-93d8b9eab885-kube-api-access-bq692\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.109207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.109274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755354bd-175b-46e6-a9bf-93d8b9eab885-logs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.109297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-combined-ca-bundle\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.109351 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq692\" (UniqueName: \"kubernetes.io/projected/755354bd-175b-46e6-a9bf-93d8b9eab885-kube-api-access-bq692\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.109549 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-public-tls-certs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.109576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data-custom\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.110041 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755354bd-175b-46e6-a9bf-93d8b9eab885-logs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.110383 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-internal-tls-certs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.115206 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data-custom\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.115562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-combined-ca-bundle\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.116207 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.123279 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-internal-tls-certs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.127195 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-public-tls-certs\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.135058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq692\" (UniqueName: \"kubernetes.io/projected/755354bd-175b-46e6-a9bf-93d8b9eab885-kube-api-access-bq692\") pod \"barbican-api-847649ff46-ldvld\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.182226 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.326157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.573926 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api-log" containerID="cri-o://e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d" gracePeriod=30 Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.574831 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api" containerID="cri-o://fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad" gracePeriod=30 Dec 05 11:31:04 crc kubenswrapper[4809]: I1205 11:31:04.779371 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-847649ff46-ldvld"] Dec 05 11:31:04 crc kubenswrapper[4809]: E1205 11:31:04.941317 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aa5170f_d144_4731_adba_c25a63f36ac7.slice/crio-fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aa5170f_d144_4731_adba_c25a63f36ac7.slice/crio-conmon-fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad.scope\": RecentStats: unable to find data in memory cache]" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.085407 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.127473 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data-custom\") pod \"9aa5170f-d144-4731-adba-c25a63f36ac7\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.130126 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-combined-ca-bundle\") pod \"9aa5170f-d144-4731-adba-c25a63f36ac7\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.130276 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-scripts\") pod \"9aa5170f-d144-4731-adba-c25a63f36ac7\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.130411 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9aa5170f-d144-4731-adba-c25a63f36ac7-etc-machine-id\") pod \"9aa5170f-d144-4731-adba-c25a63f36ac7\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.130523 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa5170f-d144-4731-adba-c25a63f36ac7-logs\") pod \"9aa5170f-d144-4731-adba-c25a63f36ac7\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.130693 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data\") pod \"9aa5170f-d144-4731-adba-c25a63f36ac7\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.130848 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7xtq\" (UniqueName: \"kubernetes.io/projected/9aa5170f-d144-4731-adba-c25a63f36ac7-kube-api-access-f7xtq\") pod \"9aa5170f-d144-4731-adba-c25a63f36ac7\" (UID: \"9aa5170f-d144-4731-adba-c25a63f36ac7\") " Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.131914 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9aa5170f-d144-4731-adba-c25a63f36ac7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9aa5170f-d144-4731-adba-c25a63f36ac7" (UID: "9aa5170f-d144-4731-adba-c25a63f36ac7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.133058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa5170f-d144-4731-adba-c25a63f36ac7-logs" (OuterVolumeSpecName: "logs") pod "9aa5170f-d144-4731-adba-c25a63f36ac7" (UID: "9aa5170f-d144-4731-adba-c25a63f36ac7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.135036 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9aa5170f-d144-4731-adba-c25a63f36ac7" (UID: "9aa5170f-d144-4731-adba-c25a63f36ac7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.137798 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-scripts" (OuterVolumeSpecName: "scripts") pod "9aa5170f-d144-4731-adba-c25a63f36ac7" (UID: "9aa5170f-d144-4731-adba-c25a63f36ac7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.138060 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa5170f-d144-4731-adba-c25a63f36ac7-kube-api-access-f7xtq" (OuterVolumeSpecName: "kube-api-access-f7xtq") pod "9aa5170f-d144-4731-adba-c25a63f36ac7" (UID: "9aa5170f-d144-4731-adba-c25a63f36ac7"). InnerVolumeSpecName "kube-api-access-f7xtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.169929 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9aa5170f-d144-4731-adba-c25a63f36ac7" (UID: "9aa5170f-d144-4731-adba-c25a63f36ac7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.204963 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data" (OuterVolumeSpecName: "config-data") pod "9aa5170f-d144-4731-adba-c25a63f36ac7" (UID: "9aa5170f-d144-4731-adba-c25a63f36ac7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.233412 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7xtq\" (UniqueName: \"kubernetes.io/projected/9aa5170f-d144-4731-adba-c25a63f36ac7-kube-api-access-f7xtq\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.233458 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.233475 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.233983 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.234026 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9aa5170f-d144-4731-adba-c25a63f36ac7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.234047 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa5170f-d144-4731-adba-c25a63f36ac7-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.234063 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa5170f-d144-4731-adba-c25a63f36ac7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.587539 4809 generic.go:334] "Generic (PLEG): container finished" podID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerID="fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad" exitCode=0 Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.587846 4809 generic.go:334] "Generic (PLEG): container finished" podID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerID="e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d" exitCode=143 Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.587686 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9aa5170f-d144-4731-adba-c25a63f36ac7","Type":"ContainerDied","Data":"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad"} Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.587941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9aa5170f-d144-4731-adba-c25a63f36ac7","Type":"ContainerDied","Data":"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d"} Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.587955 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9aa5170f-d144-4731-adba-c25a63f36ac7","Type":"ContainerDied","Data":"c5863d3c54b16a3daf7a216ae8123ecd8cf7328974446cb6a158add817d88bf4"} Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.587970 4809 scope.go:117] "RemoveContainer" containerID="fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.587764 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.592694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847649ff46-ldvld" event={"ID":"755354bd-175b-46e6-a9bf-93d8b9eab885","Type":"ContainerStarted","Data":"f75988c31f2d6e2321aedeaffd4c2a8dcc59f31b1bcb9bc5bb4646d960e0a43b"} Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.592727 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847649ff46-ldvld" event={"ID":"755354bd-175b-46e6-a9bf-93d8b9eab885","Type":"ContainerStarted","Data":"9c770678dae206680aca70e609b6ceca94a601f7c7b778f6717417cdf5387456"} Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.592737 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847649ff46-ldvld" event={"ID":"755354bd-175b-46e6-a9bf-93d8b9eab885","Type":"ContainerStarted","Data":"57f3fc6daee657204b26fd421cd7bad22af88f4ff2835b2839037129ffcda916"} Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.592906 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.606574 4809 scope.go:117] "RemoveContainer" containerID="e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.626458 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-847649ff46-ldvld" podStartSLOduration=2.626396831 podStartE2EDuration="2.626396831s" podCreationTimestamp="2025-12-05 11:31:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:05.613723399 +0000 UTC m=+1361.004699967" watchObservedRunningTime="2025-12-05 11:31:05.626396831 +0000 UTC m=+1361.017373389" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.632586 4809 scope.go:117] "RemoveContainer" containerID="fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad" Dec 05 11:31:05 crc kubenswrapper[4809]: E1205 11:31:05.636083 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad\": container with ID starting with fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad not found: ID does not exist" containerID="fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.636121 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad"} err="failed to get container status \"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad\": rpc error: code = NotFound desc = could not find container \"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad\": container with ID starting with fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad not found: ID does not exist" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.636143 4809 scope.go:117] "RemoveContainer" containerID="e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d" Dec 05 11:31:05 crc kubenswrapper[4809]: E1205 11:31:05.636608 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d\": container with ID starting with e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d not found: ID does not exist" containerID="e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.636660 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d"} err="failed to get container status \"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d\": rpc error: code = NotFound desc = could not find container \"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d\": container with ID starting with e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d not found: ID does not exist" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.636682 4809 scope.go:117] "RemoveContainer" containerID="fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.637522 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad"} err="failed to get container status \"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad\": rpc error: code = NotFound desc = could not find container \"fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad\": container with ID starting with fab1c3e2b094f865e3d9689236441454c844f596f7956e56364a7fcb4f4e01ad not found: ID does not exist" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.637551 4809 scope.go:117] "RemoveContainer" containerID="e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.637942 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d"} err="failed to get container status \"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d\": rpc error: code = NotFound desc = could not find container \"e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d\": container with ID starting with e170a1aba9da0edbc0abe14001f9b44e0d70f19506e8870f2e7287c68c86c66d not found: ID does not exist" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.645609 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.665507 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.677970 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:31:05 crc kubenswrapper[4809]: E1205 11:31:05.678357 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.678373 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api" Dec 05 11:31:05 crc kubenswrapper[4809]: E1205 11:31:05.678387 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api-log" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.678393 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api-log" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.678577 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api-log" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.678601 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" containerName="cinder-api" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.679708 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.687116 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.687280 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.687398 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.706694 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743547 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743596 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/846a6cff-41b5-4b49-8569-8c757669fe7f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743649 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743681 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743710 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmhs\" (UniqueName: \"kubernetes.io/projected/846a6cff-41b5-4b49-8569-8c757669fe7f-kube-api-access-8cmhs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743774 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743826 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data-custom\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743852 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846a6cff-41b5-4b49-8569-8c757669fe7f-logs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.743894 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845323 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845389 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data-custom\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845413 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846a6cff-41b5-4b49-8569-8c757669fe7f-logs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845652 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/846a6cff-41b5-4b49-8569-8c757669fe7f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845679 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845724 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmhs\" (UniqueName: \"kubernetes.io/projected/846a6cff-41b5-4b49-8569-8c757669fe7f-kube-api-access-8cmhs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.845887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/846a6cff-41b5-4b49-8569-8c757669fe7f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.846688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846a6cff-41b5-4b49-8569-8c757669fe7f-logs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.849831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.851052 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.851660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.852687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data-custom\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.854498 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.855436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:05 crc kubenswrapper[4809]: I1205 11:31:05.863817 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmhs\" (UniqueName: \"kubernetes.io/projected/846a6cff-41b5-4b49-8569-8c757669fe7f-kube-api-access-8cmhs\") pod \"cinder-api-0\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " pod="openstack/cinder-api-0" Dec 05 11:31:06 crc kubenswrapper[4809]: I1205 11:31:06.011193 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:31:06 crc kubenswrapper[4809]: I1205 11:31:06.515746 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:31:06 crc kubenswrapper[4809]: I1205 11:31:06.620431 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"846a6cff-41b5-4b49-8569-8c757669fe7f","Type":"ContainerStarted","Data":"7ed98d1e5c7dd98eea76484331349f5af670b8bd7119f23893508c5a1a79dc81"} Dec 05 11:31:06 crc kubenswrapper[4809]: I1205 11:31:06.620826 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:06 crc kubenswrapper[4809]: I1205 11:31:06.883649 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa5170f-d144-4731-adba-c25a63f36ac7" path="/var/lib/kubelet/pods/9aa5170f-d144-4731-adba-c25a63f36ac7/volumes" Dec 05 11:31:07 crc kubenswrapper[4809]: I1205 11:31:07.636509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"846a6cff-41b5-4b49-8569-8c757669fe7f","Type":"ContainerStarted","Data":"fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2"} Dec 05 11:31:08 crc kubenswrapper[4809]: I1205 11:31:08.507232 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:31:08 crc kubenswrapper[4809]: I1205 11:31:08.621849 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:31:08 crc kubenswrapper[4809]: I1205 11:31:08.668922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"846a6cff-41b5-4b49-8569-8c757669fe7f","Type":"ContainerStarted","Data":"ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8"} Dec 05 11:31:08 crc kubenswrapper[4809]: I1205 11:31:08.671815 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 11:31:08 crc kubenswrapper[4809]: I1205 11:31:08.767169 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.767144847 podStartE2EDuration="3.767144847s" podCreationTimestamp="2025-12-05 11:31:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:08.705323375 +0000 UTC m=+1364.096299943" watchObservedRunningTime="2025-12-05 11:31:08.767144847 +0000 UTC m=+1364.158121405" Dec 05 11:31:08 crc kubenswrapper[4809]: I1205 11:31:08.847207 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:31:08 crc kubenswrapper[4809]: I1205 11:31:08.994295 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.091914 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.166325 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-rxkmx"] Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.166704 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" podUID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerName="dnsmasq-dns" containerID="cri-o://fb2ba8a351288b4e9e91e7356e1bca3ff3d3d9546db734be03367086d01d5f39" gracePeriod=10 Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.455215 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.542114 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.677257 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerID="fb2ba8a351288b4e9e91e7356e1bca3ff3d3d9546db734be03367086d01d5f39" exitCode=0 Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.677404 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" event={"ID":"5ec92f75-6a0d-43b1-8e4e-44be2a248295","Type":"ContainerDied","Data":"fb2ba8a351288b4e9e91e7356e1bca3ff3d3d9546db734be03367086d01d5f39"} Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.677465 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" event={"ID":"5ec92f75-6a0d-43b1-8e4e-44be2a248295","Type":"ContainerDied","Data":"51ec5e6fd05a319f6631d6c2eb43b1bb613016cd117dbc3f0fcc5309c981bcf5"} Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.677469 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="cinder-scheduler" containerID="cri-o://f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5" gracePeriod=30 Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.677476 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51ec5e6fd05a319f6631d6c2eb43b1bb613016cd117dbc3f0fcc5309c981bcf5" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.677540 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="probe" containerID="cri-o://bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d" gracePeriod=30 Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.700680 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.742538 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7w88\" (UniqueName: \"kubernetes.io/projected/5ec92f75-6a0d-43b1-8e4e-44be2a248295-kube-api-access-t7w88\") pod \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.742662 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-sb\") pod \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.742752 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-config\") pod \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.742802 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-swift-storage-0\") pod \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.742897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-svc\") pod \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.742973 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-nb\") pod \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\" (UID: \"5ec92f75-6a0d-43b1-8e4e-44be2a248295\") " Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.759924 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec92f75-6a0d-43b1-8e4e-44be2a248295-kube-api-access-t7w88" (OuterVolumeSpecName: "kube-api-access-t7w88") pod "5ec92f75-6a0d-43b1-8e4e-44be2a248295" (UID: "5ec92f75-6a0d-43b1-8e4e-44be2a248295"). InnerVolumeSpecName "kube-api-access-t7w88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.845317 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7w88\" (UniqueName: \"kubernetes.io/projected/5ec92f75-6a0d-43b1-8e4e-44be2a248295-kube-api-access-t7w88\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.858562 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ec92f75-6a0d-43b1-8e4e-44be2a248295" (UID: "5ec92f75-6a0d-43b1-8e4e-44be2a248295"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.883056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5ec92f75-6a0d-43b1-8e4e-44be2a248295" (UID: "5ec92f75-6a0d-43b1-8e4e-44be2a248295"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.892938 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5ec92f75-6a0d-43b1-8e4e-44be2a248295" (UID: "5ec92f75-6a0d-43b1-8e4e-44be2a248295"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.905205 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-config" (OuterVolumeSpecName: "config") pod "5ec92f75-6a0d-43b1-8e4e-44be2a248295" (UID: "5ec92f75-6a0d-43b1-8e4e-44be2a248295"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.911465 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5ec92f75-6a0d-43b1-8e4e-44be2a248295" (UID: "5ec92f75-6a0d-43b1-8e4e-44be2a248295"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.946981 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.947016 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.947026 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.947034 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:09 crc kubenswrapper[4809]: I1205 11:31:09.947043 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ec92f75-6a0d-43b1-8e4e-44be2a248295-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:10 crc kubenswrapper[4809]: I1205 11:31:10.409961 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:31:10 crc kubenswrapper[4809]: I1205 11:31:10.688909 4809 generic.go:334] "Generic (PLEG): container finished" podID="a2557186-1276-41f5-99af-571e2db4c44e" containerID="bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d" exitCode=0 Dec 05 11:31:10 crc kubenswrapper[4809]: I1205 11:31:10.689004 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-rxkmx" Dec 05 11:31:10 crc kubenswrapper[4809]: I1205 11:31:10.688998 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a2557186-1276-41f5-99af-571e2db4c44e","Type":"ContainerDied","Data":"bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d"} Dec 05 11:31:10 crc kubenswrapper[4809]: I1205 11:31:10.728316 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-rxkmx"] Dec 05 11:31:10 crc kubenswrapper[4809]: I1205 11:31:10.742007 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-rxkmx"] Dec 05 11:31:10 crc kubenswrapper[4809]: I1205 11:31:10.883245 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" path="/var/lib/kubelet/pods/5ec92f75-6a0d-43b1-8e4e-44be2a248295/volumes" Dec 05 11:31:11 crc kubenswrapper[4809]: I1205 11:31:11.308813 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:11 crc kubenswrapper[4809]: I1205 11:31:11.562897 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.243753 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 11:31:12 crc kubenswrapper[4809]: E1205 11:31:12.244318 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerName="init" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.244336 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerName="init" Dec 05 11:31:12 crc kubenswrapper[4809]: E1205 11:31:12.244360 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerName="dnsmasq-dns" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.244368 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerName="dnsmasq-dns" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.244592 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec92f75-6a0d-43b1-8e4e-44be2a248295" containerName="dnsmasq-dns" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.245591 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.248395 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.248733 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-wg5m4" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.248989 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.266603 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.298051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.298106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rndwp\" (UniqueName: \"kubernetes.io/projected/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-kube-api-access-rndwp\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.298264 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.298322 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.400183 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.400565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.400759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.400882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rndwp\" (UniqueName: \"kubernetes.io/projected/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-kube-api-access-rndwp\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.401205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.407083 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.410675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.431379 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rndwp\" (UniqueName: \"kubernetes.io/projected/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-kube-api-access-rndwp\") pod \"openstackclient\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.515897 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.516880 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.530904 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.568443 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.569862 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.576214 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.604672 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqsln\" (UniqueName: \"kubernetes.io/projected/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-kube-api-access-hqsln\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.604730 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.604813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config-secret\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.604865 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: E1205 11:31:12.677038 4809 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 11:31:12 crc kubenswrapper[4809]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_b2b94ef0-9a1a-447e-96ba-afb0d699cb82_0(f0ca2398637569330144c0946b2d52674dd512062d608da135b2c9fedfcda1eb): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ca2398637569330144c0946b2d52674dd512062d608da135b2c9fedfcda1eb" Netns:"/var/run/netns/59074d77-34aa-4859-8d19-bb208e0d8a3d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=f0ca2398637569330144c0946b2d52674dd512062d608da135b2c9fedfcda1eb;K8S_POD_UID=b2b94ef0-9a1a-447e-96ba-afb0d699cb82" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/b2b94ef0-9a1a-447e-96ba-afb0d699cb82]: expected pod UID "b2b94ef0-9a1a-447e-96ba-afb0d699cb82" but got "5a9a6752-88c3-4c2c-a54b-2e37a953fd50" from Kube API Dec 05 11:31:12 crc kubenswrapper[4809]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 11:31:12 crc kubenswrapper[4809]: > Dec 05 11:31:12 crc kubenswrapper[4809]: E1205 11:31:12.677115 4809 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 11:31:12 crc kubenswrapper[4809]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_b2b94ef0-9a1a-447e-96ba-afb0d699cb82_0(f0ca2398637569330144c0946b2d52674dd512062d608da135b2c9fedfcda1eb): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ca2398637569330144c0946b2d52674dd512062d608da135b2c9fedfcda1eb" Netns:"/var/run/netns/59074d77-34aa-4859-8d19-bb208e0d8a3d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=f0ca2398637569330144c0946b2d52674dd512062d608da135b2c9fedfcda1eb;K8S_POD_UID=b2b94ef0-9a1a-447e-96ba-afb0d699cb82" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/b2b94ef0-9a1a-447e-96ba-afb0d699cb82]: expected pod UID "b2b94ef0-9a1a-447e-96ba-afb0d699cb82" but got "5a9a6752-88c3-4c2c-a54b-2e37a953fd50" from Kube API Dec 05 11:31:12 crc kubenswrapper[4809]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 11:31:12 crc kubenswrapper[4809]: > pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.713751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.714969 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqsln\" (UniqueName: \"kubernetes.io/projected/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-kube-api-access-hqsln\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.715163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.715437 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config-secret\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.723146 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.723737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.727331 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.729408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config-secret\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.734233 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b2b94ef0-9a1a-447e-96ba-afb0d699cb82" podUID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.736178 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqsln\" (UniqueName: \"kubernetes.io/projected/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-kube-api-access-hqsln\") pod \"openstackclient\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.737689 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.817199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config\") pod \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.817375 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-combined-ca-bundle\") pod \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.817449 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rndwp\" (UniqueName: \"kubernetes.io/projected/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-kube-api-access-rndwp\") pod \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.817515 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config-secret\") pod \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\" (UID: \"b2b94ef0-9a1a-447e-96ba-afb0d699cb82\") " Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.818022 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b2b94ef0-9a1a-447e-96ba-afb0d699cb82" (UID: "b2b94ef0-9a1a-447e-96ba-afb0d699cb82"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.818170 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.821417 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2b94ef0-9a1a-447e-96ba-afb0d699cb82" (UID: "b2b94ef0-9a1a-447e-96ba-afb0d699cb82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.821485 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-kube-api-access-rndwp" (OuterVolumeSpecName: "kube-api-access-rndwp") pod "b2b94ef0-9a1a-447e-96ba-afb0d699cb82" (UID: "b2b94ef0-9a1a-447e-96ba-afb0d699cb82"). InnerVolumeSpecName "kube-api-access-rndwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.824802 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b2b94ef0-9a1a-447e-96ba-afb0d699cb82" (UID: "b2b94ef0-9a1a-447e-96ba-afb0d699cb82"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.884370 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b94ef0-9a1a-447e-96ba-afb0d699cb82" path="/var/lib/kubelet/pods/b2b94ef0-9a1a-447e-96ba-afb0d699cb82/volumes" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.919736 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.919774 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rndwp\" (UniqueName: \"kubernetes.io/projected/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-kube-api-access-rndwp\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.919787 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2b94ef0-9a1a-447e-96ba-afb0d699cb82-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:12 crc kubenswrapper[4809]: I1205 11:31:12.954490 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.131301 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.226828 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7cb9769bdb-8tnj5"] Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.227204 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7cb9769bdb-8tnj5" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api-log" containerID="cri-o://aa267d76d96ed36f44971b8d32186525836ead41231817f2b83d510110f590cb" gracePeriod=30 Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.227431 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7cb9769bdb-8tnj5" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api" containerID="cri-o://a5c195143bfe59cebbb4f4a0366446d0f7ba8c9aa9a37e99673ab8da024eba8f" gracePeriod=30 Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.483626 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.642075 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.739799 4809 generic.go:334] "Generic (PLEG): container finished" podID="a2557186-1276-41f5-99af-571e2db4c44e" containerID="f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5" exitCode=0 Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.739884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a2557186-1276-41f5-99af-571e2db4c44e","Type":"ContainerDied","Data":"f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5"} Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.739921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a2557186-1276-41f5-99af-571e2db4c44e","Type":"ContainerDied","Data":"2ff6fc468d4b53b1e09db2073a4390bf7b073f5d675cd002d5ee8ec42cfebe84"} Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.739944 4809 scope.go:117] "RemoveContainer" containerID="bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.740106 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.742207 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-scripts\") pod \"a2557186-1276-41f5-99af-571e2db4c44e\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.742278 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pngsm\" (UniqueName: \"kubernetes.io/projected/a2557186-1276-41f5-99af-571e2db4c44e-kube-api-access-pngsm\") pod \"a2557186-1276-41f5-99af-571e2db4c44e\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.742296 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data-custom\") pod \"a2557186-1276-41f5-99af-571e2db4c44e\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.742323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2557186-1276-41f5-99af-571e2db4c44e-etc-machine-id\") pod \"a2557186-1276-41f5-99af-571e2db4c44e\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.742361 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-combined-ca-bundle\") pod \"a2557186-1276-41f5-99af-571e2db4c44e\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.742401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data\") pod \"a2557186-1276-41f5-99af-571e2db4c44e\" (UID: \"a2557186-1276-41f5-99af-571e2db4c44e\") " Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.748781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5a9a6752-88c3-4c2c-a54b-2e37a953fd50","Type":"ContainerStarted","Data":"856f7895b3ab54eb1f22d3e4564ef5fad57c4a85a55b9805d5ca893e2e01b368"} Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.748886 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2557186-1276-41f5-99af-571e2db4c44e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a2557186-1276-41f5-99af-571e2db4c44e" (UID: "a2557186-1276-41f5-99af-571e2db4c44e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.751884 4809 generic.go:334] "Generic (PLEG): container finished" podID="80cac454-9282-41a3-b376-c2beb5b0096c" containerID="aa267d76d96ed36f44971b8d32186525836ead41231817f2b83d510110f590cb" exitCode=143 Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.751940 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.751917 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2557186-1276-41f5-99af-571e2db4c44e-kube-api-access-pngsm" (OuterVolumeSpecName: "kube-api-access-pngsm") pod "a2557186-1276-41f5-99af-571e2db4c44e" (UID: "a2557186-1276-41f5-99af-571e2db4c44e"). InnerVolumeSpecName "kube-api-access-pngsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.752052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cb9769bdb-8tnj5" event={"ID":"80cac454-9282-41a3-b376-c2beb5b0096c","Type":"ContainerDied","Data":"aa267d76d96ed36f44971b8d32186525836ead41231817f2b83d510110f590cb"} Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.752770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-scripts" (OuterVolumeSpecName: "scripts") pod "a2557186-1276-41f5-99af-571e2db4c44e" (UID: "a2557186-1276-41f5-99af-571e2db4c44e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.754475 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a2557186-1276-41f5-99af-571e2db4c44e" (UID: "a2557186-1276-41f5-99af-571e2db4c44e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.813949 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b2b94ef0-9a1a-447e-96ba-afb0d699cb82" podUID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.824908 4809 scope.go:117] "RemoveContainer" containerID="f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.832169 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2557186-1276-41f5-99af-571e2db4c44e" (UID: "a2557186-1276-41f5-99af-571e2db4c44e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.845063 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.845092 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pngsm\" (UniqueName: \"kubernetes.io/projected/a2557186-1276-41f5-99af-571e2db4c44e-kube-api-access-pngsm\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.845103 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.845113 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a2557186-1276-41f5-99af-571e2db4c44e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.845121 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.877272 4809 scope.go:117] "RemoveContainer" containerID="bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d" Dec 05 11:31:13 crc kubenswrapper[4809]: E1205 11:31:13.877890 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d\": container with ID starting with bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d not found: ID does not exist" containerID="bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.877920 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d"} err="failed to get container status \"bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d\": rpc error: code = NotFound desc = could not find container \"bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d\": container with ID starting with bdf69a431c1987a1d770ad2c00faec5fc0acabb8735a9c77a700d2414fdb5a4d not found: ID does not exist" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.877938 4809 scope.go:117] "RemoveContainer" containerID="f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5" Dec 05 11:31:13 crc kubenswrapper[4809]: E1205 11:31:13.878272 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5\": container with ID starting with f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5 not found: ID does not exist" containerID="f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.878293 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5"} err="failed to get container status \"f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5\": rpc error: code = NotFound desc = could not find container \"f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5\": container with ID starting with f04d6d6c94218ce66fc4a1a3cc5ad925d7176d62c6f8e074d02c38458de8ccd5 not found: ID does not exist" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.885736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data" (OuterVolumeSpecName: "config-data") pod "a2557186-1276-41f5-99af-571e2db4c44e" (UID: "a2557186-1276-41f5-99af-571e2db4c44e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:13 crc kubenswrapper[4809]: I1205 11:31:13.947314 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2557186-1276-41f5-99af-571e2db4c44e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.046886 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.047186 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.047252 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.048014 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09f7d635e94f05ae8e8f470c73493c96f0ec6c0a4510de5411cdb6df882057b9"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.048065 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://09f7d635e94f05ae8e8f470c73493c96f0ec6c0a4510de5411cdb6df882057b9" gracePeriod=600 Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.088613 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.114397 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.123226 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:31:14 crc kubenswrapper[4809]: E1205 11:31:14.123843 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="cinder-scheduler" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.123871 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="cinder-scheduler" Dec 05 11:31:14 crc kubenswrapper[4809]: E1205 11:31:14.123894 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="probe" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.123905 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="probe" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.124175 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="probe" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.124201 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2557186-1276-41f5-99af-571e2db4c44e" containerName="cinder-scheduler" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.125440 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.130103 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.135104 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.155854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.156131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-scripts\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.156173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.156230 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d12a95a-49fa-4392-9a0d-9efd9aa28720-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.156256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbl6g\" (UniqueName: \"kubernetes.io/projected/3d12a95a-49fa-4392-9a0d-9efd9aa28720-kube-api-access-hbl6g\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.156331 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.258966 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-scripts\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.259015 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.259055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d12a95a-49fa-4392-9a0d-9efd9aa28720-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.259073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbl6g\" (UniqueName: \"kubernetes.io/projected/3d12a95a-49fa-4392-9a0d-9efd9aa28720-kube-api-access-hbl6g\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.259117 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.259149 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.260071 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d12a95a-49fa-4392-9a0d-9efd9aa28720-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.264544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.264732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.264941 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-scripts\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.279140 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.282222 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbl6g\" (UniqueName: \"kubernetes.io/projected/3d12a95a-49fa-4392-9a0d-9efd9aa28720-kube-api-access-hbl6g\") pod \"cinder-scheduler-0\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.437318 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.479172 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.515046 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-df99668dd-llrdv"] Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.519417 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-df99668dd-llrdv" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-api" containerID="cri-o://bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4" gracePeriod=30 Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.520160 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-df99668dd-llrdv" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-httpd" containerID="cri-o://ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715" gracePeriod=30 Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.784898 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="09f7d635e94f05ae8e8f470c73493c96f0ec6c0a4510de5411cdb6df882057b9" exitCode=0 Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.785329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"09f7d635e94f05ae8e8f470c73493c96f0ec6c0a4510de5411cdb6df882057b9"} Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.785368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9"} Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.785388 4809 scope.go:117] "RemoveContainer" containerID="f84b335fda3d018d7628fbb24eef2ea4ed19d76e6c066be3637c91567a56e36e" Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.793077 4809 generic.go:334] "Generic (PLEG): container finished" podID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerID="ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715" exitCode=0 Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.793161 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-df99668dd-llrdv" event={"ID":"782fab1a-0354-47d4-8641-cfc7baf2bdf2","Type":"ContainerDied","Data":"ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715"} Dec 05 11:31:14 crc kubenswrapper[4809]: I1205 11:31:14.890749 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2557186-1276-41f5-99af-571e2db4c44e" path="/var/lib/kubelet/pods/a2557186-1276-41f5-99af-571e2db4c44e/volumes" Dec 05 11:31:15 crc kubenswrapper[4809]: I1205 11:31:15.062424 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:31:15 crc kubenswrapper[4809]: I1205 11:31:15.823931 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3d12a95a-49fa-4392-9a0d-9efd9aa28720","Type":"ContainerStarted","Data":"ed7d88f2530268a3f4438c5034a010adcd8f5b9f881c741aaee213b005a6a326"} Dec 05 11:31:15 crc kubenswrapper[4809]: I1205 11:31:15.824267 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3d12a95a-49fa-4392-9a0d-9efd9aa28720","Type":"ContainerStarted","Data":"d7a7389b0b4ebe9335b1d9fe477e4e17ba0dda482d71572edcbd983856081ad2"} Dec 05 11:31:16 crc kubenswrapper[4809]: I1205 11:31:16.890206 4809 generic.go:334] "Generic (PLEG): container finished" podID="80cac454-9282-41a3-b376-c2beb5b0096c" containerID="a5c195143bfe59cebbb4f4a0366446d0f7ba8c9aa9a37e99673ab8da024eba8f" exitCode=0 Dec 05 11:31:16 crc kubenswrapper[4809]: I1205 11:31:16.932385 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cb9769bdb-8tnj5" event={"ID":"80cac454-9282-41a3-b376-c2beb5b0096c","Type":"ContainerDied","Data":"a5c195143bfe59cebbb4f4a0366446d0f7ba8c9aa9a37e99673ab8da024eba8f"} Dec 05 11:31:16 crc kubenswrapper[4809]: I1205 11:31:16.932415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3d12a95a-49fa-4392-9a0d-9efd9aa28720","Type":"ContainerStarted","Data":"76602c05e47629f8dd3a939796b77e7372f979cf7d26325edfe41b5b970ca71c"} Dec 05 11:31:16 crc kubenswrapper[4809]: I1205 11:31:16.955337 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.955310582 podStartE2EDuration="2.955310582s" podCreationTimestamp="2025-12-05 11:31:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:16.941707305 +0000 UTC m=+1372.332683873" watchObservedRunningTime="2025-12-05 11:31:16.955310582 +0000 UTC m=+1372.346287140" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.010030 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.021897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80cac454-9282-41a3-b376-c2beb5b0096c-logs\") pod \"80cac454-9282-41a3-b376-c2beb5b0096c\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.021957 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data-custom\") pod \"80cac454-9282-41a3-b376-c2beb5b0096c\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.021981 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-combined-ca-bundle\") pod \"80cac454-9282-41a3-b376-c2beb5b0096c\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.021995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data\") pod \"80cac454-9282-41a3-b376-c2beb5b0096c\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.022045 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwr72\" (UniqueName: \"kubernetes.io/projected/80cac454-9282-41a3-b376-c2beb5b0096c-kube-api-access-pwr72\") pod \"80cac454-9282-41a3-b376-c2beb5b0096c\" (UID: \"80cac454-9282-41a3-b376-c2beb5b0096c\") " Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.023777 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80cac454-9282-41a3-b376-c2beb5b0096c-logs" (OuterVolumeSpecName: "logs") pod "80cac454-9282-41a3-b376-c2beb5b0096c" (UID: "80cac454-9282-41a3-b376-c2beb5b0096c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.032793 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "80cac454-9282-41a3-b376-c2beb5b0096c" (UID: "80cac454-9282-41a3-b376-c2beb5b0096c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.087330 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80cac454-9282-41a3-b376-c2beb5b0096c-kube-api-access-pwr72" (OuterVolumeSpecName: "kube-api-access-pwr72") pod "80cac454-9282-41a3-b376-c2beb5b0096c" (UID: "80cac454-9282-41a3-b376-c2beb5b0096c"). InnerVolumeSpecName "kube-api-access-pwr72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.099475 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80cac454-9282-41a3-b376-c2beb5b0096c" (UID: "80cac454-9282-41a3-b376-c2beb5b0096c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.124505 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwr72\" (UniqueName: \"kubernetes.io/projected/80cac454-9282-41a3-b376-c2beb5b0096c-kube-api-access-pwr72\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.124783 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80cac454-9282-41a3-b376-c2beb5b0096c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.124871 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.124943 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.164344 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data" (OuterVolumeSpecName: "config-data") pod "80cac454-9282-41a3-b376-c2beb5b0096c" (UID: "80cac454-9282-41a3-b376-c2beb5b0096c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.227482 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80cac454-9282-41a3-b376-c2beb5b0096c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.941874 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cb9769bdb-8tnj5" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.941873 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cb9769bdb-8tnj5" event={"ID":"80cac454-9282-41a3-b376-c2beb5b0096c","Type":"ContainerDied","Data":"c68105d6e46325de7ac1d4835c42b9b2d894c8cd41bf640e01934c67c5ba2b78"} Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.942278 4809 scope.go:117] "RemoveContainer" containerID="a5c195143bfe59cebbb4f4a0366446d0f7ba8c9aa9a37e99673ab8da024eba8f" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.977580 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7cb9769bdb-8tnj5"] Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.977867 4809 scope.go:117] "RemoveContainer" containerID="aa267d76d96ed36f44971b8d32186525836ead41231817f2b83d510110f590cb" Dec 05 11:31:17 crc kubenswrapper[4809]: I1205 11:31:17.987535 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7cb9769bdb-8tnj5"] Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.317973 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.626089 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-78555c6b85-lv8k2"] Dec 05 11:31:18 crc kubenswrapper[4809]: E1205 11:31:18.626880 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api-log" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.626900 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api-log" Dec 05 11:31:18 crc kubenswrapper[4809]: E1205 11:31:18.626933 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.626941 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.627154 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api-log" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.627182 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" containerName="barbican-api" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.628386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.639422 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.639791 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.639971 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.644481 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78555c6b85-lv8k2"] Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667074 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-etc-swift\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667151 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk7f6\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-kube-api-access-jk7f6\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-combined-ca-bundle\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-internal-tls-certs\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667251 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-run-httpd\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-config-data\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667328 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-public-tls-certs\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.667377 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-log-httpd\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769612 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-log-httpd\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-etc-swift\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769811 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk7f6\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-kube-api-access-jk7f6\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-combined-ca-bundle\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-internal-tls-certs\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769907 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-run-httpd\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-config-data\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.769980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-public-tls-certs\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.771515 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-run-httpd\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.772058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-log-httpd\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.783884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-public-tls-certs\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.783934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-combined-ca-bundle\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.784570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-internal-tls-certs\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.784718 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-etc-swift\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.788524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-config-data\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.820541 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk7f6\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-kube-api-access-jk7f6\") pod \"swift-proxy-78555c6b85-lv8k2\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.891595 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80cac454-9282-41a3-b376-c2beb5b0096c" path="/var/lib/kubelet/pods/80cac454-9282-41a3-b376-c2beb5b0096c/volumes" Dec 05 11:31:18 crc kubenswrapper[4809]: I1205 11:31:18.983131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.480023 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.598265 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.608599 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78555c6b85-lv8k2"] Dec 05 11:31:19 crc kubenswrapper[4809]: W1205 11:31:19.613245 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafdb0190_0a3d_482e_acbc_5d40e7999345.slice/crio-c44c2c44654af38c08a5053cd17a9125148df8349fc3d004eff148bba529db6a WatchSource:0}: Error finding container c44c2c44654af38c08a5053cd17a9125148df8349fc3d004eff148bba529db6a: Status 404 returned error can't find the container with id c44c2c44654af38c08a5053cd17a9125148df8349fc3d004eff148bba529db6a Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.688759 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-config\") pod \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.689051 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-combined-ca-bundle\") pod \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.689312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9klnj\" (UniqueName: \"kubernetes.io/projected/782fab1a-0354-47d4-8641-cfc7baf2bdf2-kube-api-access-9klnj\") pod \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.689481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-ovndb-tls-certs\") pod \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.689719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-httpd-config\") pod \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\" (UID: \"782fab1a-0354-47d4-8641-cfc7baf2bdf2\") " Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.693754 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782fab1a-0354-47d4-8641-cfc7baf2bdf2-kube-api-access-9klnj" (OuterVolumeSpecName: "kube-api-access-9klnj") pod "782fab1a-0354-47d4-8641-cfc7baf2bdf2" (UID: "782fab1a-0354-47d4-8641-cfc7baf2bdf2"). InnerVolumeSpecName "kube-api-access-9klnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.694030 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "782fab1a-0354-47d4-8641-cfc7baf2bdf2" (UID: "782fab1a-0354-47d4-8641-cfc7baf2bdf2"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.740619 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-config" (OuterVolumeSpecName: "config") pod "782fab1a-0354-47d4-8641-cfc7baf2bdf2" (UID: "782fab1a-0354-47d4-8641-cfc7baf2bdf2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.754835 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "782fab1a-0354-47d4-8641-cfc7baf2bdf2" (UID: "782fab1a-0354-47d4-8641-cfc7baf2bdf2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.791715 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.791752 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9klnj\" (UniqueName: \"kubernetes.io/projected/782fab1a-0354-47d4-8641-cfc7baf2bdf2-kube-api-access-9klnj\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.791768 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.791781 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.820741 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "782fab1a-0354-47d4-8641-cfc7baf2bdf2" (UID: "782fab1a-0354-47d4-8641-cfc7baf2bdf2"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.893535 4809 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/782fab1a-0354-47d4-8641-cfc7baf2bdf2-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:19 crc kubenswrapper[4809]: I1205 11:31:19.984515 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78555c6b85-lv8k2" event={"ID":"afdb0190-0a3d-482e-acbc-5d40e7999345","Type":"ContainerStarted","Data":"c44c2c44654af38c08a5053cd17a9125148df8349fc3d004eff148bba529db6a"} Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.018375 4809 generic.go:334] "Generic (PLEG): container finished" podID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerID="bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4" exitCode=0 Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.018856 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-df99668dd-llrdv" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.019828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-df99668dd-llrdv" event={"ID":"782fab1a-0354-47d4-8641-cfc7baf2bdf2","Type":"ContainerDied","Data":"bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4"} Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.019905 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-df99668dd-llrdv" event={"ID":"782fab1a-0354-47d4-8641-cfc7baf2bdf2","Type":"ContainerDied","Data":"6d53b9be2dca41fdb5b029a89d2e7eeae2a6a51c0d00c5c19efbc26b18e0cad2"} Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.019937 4809 scope.go:117] "RemoveContainer" containerID="ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.127878 4809 scope.go:117] "RemoveContainer" containerID="bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.131555 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-df99668dd-llrdv"] Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.142036 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-df99668dd-llrdv"] Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.181755 4809 scope.go:117] "RemoveContainer" containerID="ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715" Dec 05 11:31:20 crc kubenswrapper[4809]: E1205 11:31:20.182407 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715\": container with ID starting with ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715 not found: ID does not exist" containerID="ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.182447 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715"} err="failed to get container status \"ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715\": rpc error: code = NotFound desc = could not find container \"ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715\": container with ID starting with ce8c9f1f9b8083268b35548653e035ae3071c272aeeddef238bf6a7adff81715 not found: ID does not exist" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.182474 4809 scope.go:117] "RemoveContainer" containerID="bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4" Dec 05 11:31:20 crc kubenswrapper[4809]: E1205 11:31:20.182949 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4\": container with ID starting with bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4 not found: ID does not exist" containerID="bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.183023 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4"} err="failed to get container status \"bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4\": rpc error: code = NotFound desc = could not find container \"bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4\": container with ID starting with bd6a1ca5e4e3d02febc3bbb7684c48e9418e984d5c6f107a6f4f70c127815fe4 not found: ID does not exist" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.588578 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.588875 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-central-agent" containerID="cri-o://3eceec205c19750415c3ff9e7f89e9aeb54d239bff3973eb107f4253f2992346" gracePeriod=30 Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.588925 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="proxy-httpd" containerID="cri-o://f57b5e787af7f2f1fb584535e6c2918eab3d4b7aed79704f311d5c22e1ee9e22" gracePeriod=30 Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.588983 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="sg-core" containerID="cri-o://a7544ff88a950bd54ea465532a74b580658be779670e8973cb5b65032dcf574f" gracePeriod=30 Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.589015 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-notification-agent" containerID="cri-o://e24f45701e293ee5b251a98a42d425ad6a6a5c84c03dc03ad9d44e980855289d" gracePeriod=30 Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.595250 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 11:31:20 crc kubenswrapper[4809]: I1205 11:31:20.884762 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" path="/var/lib/kubelet/pods/782fab1a-0354-47d4-8641-cfc7baf2bdf2/volumes" Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.033158 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerID="f57b5e787af7f2f1fb584535e6c2918eab3d4b7aed79704f311d5c22e1ee9e22" exitCode=0 Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.033204 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerID="a7544ff88a950bd54ea465532a74b580658be779670e8973cb5b65032dcf574f" exitCode=2 Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.033214 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerID="3eceec205c19750415c3ff9e7f89e9aeb54d239bff3973eb107f4253f2992346" exitCode=0 Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.033280 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerDied","Data":"f57b5e787af7f2f1fb584535e6c2918eab3d4b7aed79704f311d5c22e1ee9e22"} Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.033355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerDied","Data":"a7544ff88a950bd54ea465532a74b580658be779670e8973cb5b65032dcf574f"} Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.033371 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerDied","Data":"3eceec205c19750415c3ff9e7f89e9aeb54d239bff3973eb107f4253f2992346"} Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.036912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78555c6b85-lv8k2" event={"ID":"afdb0190-0a3d-482e-acbc-5d40e7999345","Type":"ContainerStarted","Data":"a3de22e1ff3f7898f9fee8f16d5df757205e16ec84e88652001b2d777b01c877"} Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.036953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78555c6b85-lv8k2" event={"ID":"afdb0190-0a3d-482e-acbc-5d40e7999345","Type":"ContainerStarted","Data":"0c7d459938d65d4994155346009222c7ccaf5872fe4d76d2c6e4cb10a6a0e64d"} Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.037192 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:21 crc kubenswrapper[4809]: I1205 11:31:21.065221 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-78555c6b85-lv8k2" podStartSLOduration=3.065194059 podStartE2EDuration="3.065194059s" podCreationTimestamp="2025-12-05 11:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:21.054460077 +0000 UTC m=+1376.445436635" watchObservedRunningTime="2025-12-05 11:31:21.065194059 +0000 UTC m=+1376.456170617" Dec 05 11:31:22 crc kubenswrapper[4809]: I1205 11:31:22.046736 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:23 crc kubenswrapper[4809]: I1205 11:31:23.062880 4809 generic.go:334] "Generic (PLEG): container finished" podID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerID="e24f45701e293ee5b251a98a42d425ad6a6a5c84c03dc03ad9d44e980855289d" exitCode=0 Dec 05 11:31:23 crc kubenswrapper[4809]: I1205 11:31:23.062927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerDied","Data":"e24f45701e293ee5b251a98a42d425ad6a6a5c84c03dc03ad9d44e980855289d"} Dec 05 11:31:23 crc kubenswrapper[4809]: I1205 11:31:23.672047 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.155:3000/\": dial tcp 10.217.0.155:3000: connect: connection refused" Dec 05 11:31:24 crc kubenswrapper[4809]: I1205 11:31:24.760522 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 11:31:25 crc kubenswrapper[4809]: I1205 11:31:25.530342 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:31:25 crc kubenswrapper[4809]: I1205 11:31:25.531814 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="bcee5292-e219-4b51-871e-a283dc5cb231" containerName="kube-state-metrics" containerID="cri-o://62fac1898d75b05fbf7bcbdeb2f3a3ddc669467f4255468d71872dab180ef4b4" gracePeriod=30 Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.092914 4809 generic.go:334] "Generic (PLEG): container finished" podID="bcee5292-e219-4b51-871e-a283dc5cb231" containerID="62fac1898d75b05fbf7bcbdeb2f3a3ddc669467f4255468d71872dab180ef4b4" exitCode=2 Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.092973 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bcee5292-e219-4b51-871e-a283dc5cb231","Type":"ContainerDied","Data":"62fac1898d75b05fbf7bcbdeb2f3a3ddc669467f4255468d71872dab180ef4b4"} Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.727697 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.777863 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.845780 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-run-httpd\") pod \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.845880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-config-data\") pod \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.845923 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtfxv\" (UniqueName: \"kubernetes.io/projected/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-kube-api-access-rtfxv\") pod \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.845966 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-sg-core-conf-yaml\") pod \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.845993 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-log-httpd\") pod \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.846016 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-scripts\") pod \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.846057 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-combined-ca-bundle\") pod \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\" (UID: \"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.846866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" (UID: "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.847916 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" (UID: "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.851799 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-scripts" (OuterVolumeSpecName: "scripts") pod "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" (UID: "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.851920 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-kube-api-access-rtfxv" (OuterVolumeSpecName: "kube-api-access-rtfxv") pod "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" (UID: "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a"). InnerVolumeSpecName "kube-api-access-rtfxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.911261 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" (UID: "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.935974 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" (UID: "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.947487 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggrjx\" (UniqueName: \"kubernetes.io/projected/bcee5292-e219-4b51-871e-a283dc5cb231-kube-api-access-ggrjx\") pod \"bcee5292-e219-4b51-871e-a283dc5cb231\" (UID: \"bcee5292-e219-4b51-871e-a283dc5cb231\") " Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.948177 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtfxv\" (UniqueName: \"kubernetes.io/projected/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-kube-api-access-rtfxv\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.948208 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.948219 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.948228 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.948237 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.948246 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.950089 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcee5292-e219-4b51-871e-a283dc5cb231-kube-api-access-ggrjx" (OuterVolumeSpecName: "kube-api-access-ggrjx") pod "bcee5292-e219-4b51-871e-a283dc5cb231" (UID: "bcee5292-e219-4b51-871e-a283dc5cb231"). InnerVolumeSpecName "kube-api-access-ggrjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:26 crc kubenswrapper[4809]: I1205 11:31:26.957103 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-config-data" (OuterVolumeSpecName: "config-data") pod "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" (UID: "7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.049616 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.049655 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggrjx\" (UniqueName: \"kubernetes.io/projected/bcee5292-e219-4b51-871e-a283dc5cb231-kube-api-access-ggrjx\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.105927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5a9a6752-88c3-4c2c-a54b-2e37a953fd50","Type":"ContainerStarted","Data":"719fed1b9e606f1acf047af8258117088542a8ab9a3b33d882683eb029e590fd"} Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.108525 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bcee5292-e219-4b51-871e-a283dc5cb231","Type":"ContainerDied","Data":"c24383c5e8e7ea5ddf62d3c90c226f0f8352b4f73f54415b7ab5727885be4da2"} Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.108570 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.108602 4809 scope.go:117] "RemoveContainer" containerID="62fac1898d75b05fbf7bcbdeb2f3a3ddc669467f4255468d71872dab180ef4b4" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.110977 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a","Type":"ContainerDied","Data":"7ce99acbeeb400a512c03176032cc674e88522f96065b1501de93e9484899ec3"} Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.111104 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.127351 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.170918447 podStartE2EDuration="15.127327604s" podCreationTimestamp="2025-12-05 11:31:12 +0000 UTC" firstStartedPulling="2025-12-05 11:31:13.490853171 +0000 UTC m=+1368.881829729" lastFinishedPulling="2025-12-05 11:31:26.447262328 +0000 UTC m=+1381.838238886" observedRunningTime="2025-12-05 11:31:27.11918358 +0000 UTC m=+1382.510160148" watchObservedRunningTime="2025-12-05 11:31:27.127327604 +0000 UTC m=+1382.518304162" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.137511 4809 scope.go:117] "RemoveContainer" containerID="f57b5e787af7f2f1fb584535e6c2918eab3d4b7aed79704f311d5c22e1ee9e22" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.168694 4809 scope.go:117] "RemoveContainer" containerID="a7544ff88a950bd54ea465532a74b580658be779670e8973cb5b65032dcf574f" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.171153 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.192555 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.215149 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.215746 4809 scope.go:117] "RemoveContainer" containerID="e24f45701e293ee5b251a98a42d425ad6a6a5c84c03dc03ad9d44e980855289d" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.227326 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.244775 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: E1205 11:31:27.245303 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="sg-core" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.245330 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="sg-core" Dec 05 11:31:27 crc kubenswrapper[4809]: E1205 11:31:27.245347 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-central-agent" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.245355 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-central-agent" Dec 05 11:31:27 crc kubenswrapper[4809]: E1205 11:31:27.245367 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-notification-agent" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.245377 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-notification-agent" Dec 05 11:31:27 crc kubenswrapper[4809]: E1205 11:31:27.245606 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="proxy-httpd" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.245617 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="proxy-httpd" Dec 05 11:31:27 crc kubenswrapper[4809]: E1205 11:31:27.245656 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-httpd" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.245668 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-httpd" Dec 05 11:31:27 crc kubenswrapper[4809]: E1205 11:31:27.245684 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-api" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.245694 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-api" Dec 05 11:31:27 crc kubenswrapper[4809]: E1205 11:31:27.245715 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee5292-e219-4b51-871e-a283dc5cb231" containerName="kube-state-metrics" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.245722 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee5292-e219-4b51-871e-a283dc5cb231" containerName="kube-state-metrics" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.246289 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcee5292-e219-4b51-871e-a283dc5cb231" containerName="kube-state-metrics" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.246320 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-httpd" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.246342 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="782fab1a-0354-47d4-8641-cfc7baf2bdf2" containerName="neutron-api" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.246357 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="proxy-httpd" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.246377 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-central-agent" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.246387 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="sg-core" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.246397 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" containerName="ceilometer-notification-agent" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.254002 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.254212 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.259290 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4jnk8" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.260015 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.260164 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.260775 4809 scope.go:117] "RemoveContainer" containerID="3eceec205c19750415c3ff9e7f89e9aeb54d239bff3973eb107f4253f2992346" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.261190 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.264014 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.267127 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.267165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.268596 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.357518 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62lp4\" (UniqueName: \"kubernetes.io/projected/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-api-access-62lp4\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.357657 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.357947 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.358106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-scripts\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460122 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-config-data\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460274 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460320 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-run-httpd\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460525 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460937 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l5qq\" (UniqueName: \"kubernetes.io/projected/50427365-4dce-49d9-8d35-09866162c6de-kube-api-access-7l5qq\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.460982 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.461026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62lp4\" (UniqueName: \"kubernetes.io/projected/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-api-access-62lp4\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.461067 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-log-httpd\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.461131 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.465589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.466802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.467417 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.486229 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62lp4\" (UniqueName: \"kubernetes.io/projected/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-api-access-62lp4\") pod \"kube-state-metrics-0\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.562318 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.562419 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-log-httpd\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.562553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-scripts\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.562933 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-log-httpd\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.563013 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-config-data\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.563056 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.563092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-run-httpd\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.563191 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l5qq\" (UniqueName: \"kubernetes.io/projected/50427365-4dce-49d9-8d35-09866162c6de-kube-api-access-7l5qq\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.564059 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-run-httpd\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.566021 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.566505 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-config-data\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.566517 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.567549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-scripts\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.581793 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l5qq\" (UniqueName: \"kubernetes.io/projected/50427365-4dce-49d9-8d35-09866162c6de-kube-api-access-7l5qq\") pod \"ceilometer-0\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.600839 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.609894 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:27 crc kubenswrapper[4809]: I1205 11:31:27.940801 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.047570 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:31:28 crc kubenswrapper[4809]: W1205 11:31:28.052106 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21e68543_4956_4fc0_86aa_9ae8484b07c3.slice/crio-3233d6ed4f63a39d0cf7c9550d125b9e7c40bf22e15894ced65db71000e62605 WatchSource:0}: Error finding container 3233d6ed4f63a39d0cf7c9550d125b9e7c40bf22e15894ced65db71000e62605: Status 404 returned error can't find the container with id 3233d6ed4f63a39d0cf7c9550d125b9e7c40bf22e15894ced65db71000e62605 Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.125234 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21e68543-4956-4fc0-86aa-9ae8484b07c3","Type":"ContainerStarted","Data":"3233d6ed4f63a39d0cf7c9550d125b9e7c40bf22e15894ced65db71000e62605"} Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.138371 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:28 crc kubenswrapper[4809]: W1205 11:31:28.142017 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50427365_4dce_49d9_8d35_09866162c6de.slice/crio-c9f0655ebdcf793c3e27fc709d02ef773ddb54e565e3b173984dbf181b003d08 WatchSource:0}: Error finding container c9f0655ebdcf793c3e27fc709d02ef773ddb54e565e3b173984dbf181b003d08: Status 404 returned error can't find the container with id c9f0655ebdcf793c3e27fc709d02ef773ddb54e565e3b173984dbf181b003d08 Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.280853 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-h8l7j"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.282571 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.296846 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-h8l7j"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.387393 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6kb6p"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.388830 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.407064 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6kb6p"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.478997 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhsks\" (UniqueName: \"kubernetes.io/projected/2b9c9e47-6fdc-4f5b-845f-503cf803036a-kube-api-access-xhsks\") pod \"nova-api-db-create-h8l7j\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.479077 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b9c9e47-6fdc-4f5b-845f-503cf803036a-operator-scripts\") pod \"nova-api-db-create-h8l7j\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.494852 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-293e-account-create-update-82lc7"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.496168 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.500027 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.505197 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-293e-account-create-update-82lc7"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.592041 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhsks\" (UniqueName: \"kubernetes.io/projected/2b9c9e47-6fdc-4f5b-845f-503cf803036a-kube-api-access-xhsks\") pod \"nova-api-db-create-h8l7j\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.592102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkxs7\" (UniqueName: \"kubernetes.io/projected/6c316081-5125-4ba2-8a5d-87174b3066d2-kube-api-access-xkxs7\") pod \"nova-cell0-db-create-6kb6p\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.592147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c316081-5125-4ba2-8a5d-87174b3066d2-operator-scripts\") pod \"nova-cell0-db-create-6kb6p\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.592181 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b9c9e47-6fdc-4f5b-845f-503cf803036a-operator-scripts\") pod \"nova-api-db-create-h8l7j\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.593062 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b9c9e47-6fdc-4f5b-845f-503cf803036a-operator-scripts\") pod \"nova-api-db-create-h8l7j\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.594856 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ktb2f"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.600383 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.617578 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhsks\" (UniqueName: \"kubernetes.io/projected/2b9c9e47-6fdc-4f5b-845f-503cf803036a-kube-api-access-xhsks\") pod \"nova-api-db-create-h8l7j\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.617670 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ktb2f"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.686196 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0f47-account-create-update-f7kxx"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.687396 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.689955 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.694112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-operator-scripts\") pod \"nova-api-293e-account-create-update-82lc7\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.694220 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rh2g\" (UniqueName: \"kubernetes.io/projected/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-kube-api-access-2rh2g\") pod \"nova-api-293e-account-create-update-82lc7\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.694308 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkxs7\" (UniqueName: \"kubernetes.io/projected/6c316081-5125-4ba2-8a5d-87174b3066d2-kube-api-access-xkxs7\") pod \"nova-cell0-db-create-6kb6p\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.694333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c316081-5125-4ba2-8a5d-87174b3066d2-operator-scripts\") pod \"nova-cell0-db-create-6kb6p\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.695048 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c316081-5125-4ba2-8a5d-87174b3066d2-operator-scripts\") pod \"nova-cell0-db-create-6kb6p\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.710914 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0f47-account-create-update-f7kxx"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.715670 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkxs7\" (UniqueName: \"kubernetes.io/projected/6c316081-5125-4ba2-8a5d-87174b3066d2-kube-api-access-xkxs7\") pod \"nova-cell0-db-create-6kb6p\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.716039 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.795796 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3931783-c543-4ec1-b91c-3b2a60c4b177-operator-scripts\") pod \"nova-cell0-0f47-account-create-update-f7kxx\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.796066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0813b74-3d43-48a9-a9af-738df2fe08ad-operator-scripts\") pod \"nova-cell1-db-create-ktb2f\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.796117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zxmt\" (UniqueName: \"kubernetes.io/projected/f0813b74-3d43-48a9-a9af-738df2fe08ad-kube-api-access-5zxmt\") pod \"nova-cell1-db-create-ktb2f\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.796134 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6mrc\" (UniqueName: \"kubernetes.io/projected/b3931783-c543-4ec1-b91c-3b2a60c4b177-kube-api-access-f6mrc\") pod \"nova-cell0-0f47-account-create-update-f7kxx\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.796163 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-operator-scripts\") pod \"nova-api-293e-account-create-update-82lc7\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.796265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rh2g\" (UniqueName: \"kubernetes.io/projected/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-kube-api-access-2rh2g\") pod \"nova-api-293e-account-create-update-82lc7\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.797392 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-operator-scripts\") pod \"nova-api-293e-account-create-update-82lc7\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.813855 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rh2g\" (UniqueName: \"kubernetes.io/projected/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-kube-api-access-2rh2g\") pod \"nova-api-293e-account-create-update-82lc7\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.897847 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3931783-c543-4ec1-b91c-3b2a60c4b177-operator-scripts\") pod \"nova-cell0-0f47-account-create-update-f7kxx\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.897909 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0813b74-3d43-48a9-a9af-738df2fe08ad-operator-scripts\") pod \"nova-cell1-db-create-ktb2f\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.898525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3931783-c543-4ec1-b91c-3b2a60c4b177-operator-scripts\") pod \"nova-cell0-0f47-account-create-update-f7kxx\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.898620 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zxmt\" (UniqueName: \"kubernetes.io/projected/f0813b74-3d43-48a9-a9af-738df2fe08ad-kube-api-access-5zxmt\") pod \"nova-cell1-db-create-ktb2f\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.900298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6mrc\" (UniqueName: \"kubernetes.io/projected/b3931783-c543-4ec1-b91c-3b2a60c4b177-kube-api-access-f6mrc\") pod \"nova-cell0-0f47-account-create-update-f7kxx\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.899124 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0813b74-3d43-48a9-a9af-738df2fe08ad-operator-scripts\") pod \"nova-cell1-db-create-ktb2f\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.902731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.918769 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a" path="/var/lib/kubelet/pods/7d31eba5-7cfa-4c9f-b4fe-32a595bfb24a/volumes" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.922352 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcee5292-e219-4b51-871e-a283dc5cb231" path="/var/lib/kubelet/pods/bcee5292-e219-4b51-871e-a283dc5cb231/volumes" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.923709 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5fdb-account-create-update-92tdn"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.927183 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5fdb-account-create-update-92tdn"] Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.927285 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.927719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6mrc\" (UniqueName: \"kubernetes.io/projected/b3931783-c543-4ec1-b91c-3b2a60c4b177-kube-api-access-f6mrc\") pod \"nova-cell0-0f47-account-create-update-f7kxx\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.928715 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zxmt\" (UniqueName: \"kubernetes.io/projected/f0813b74-3d43-48a9-a9af-738df2fe08ad-kube-api-access-5zxmt\") pod \"nova-cell1-db-create-ktb2f\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.929842 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.965159 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.998672 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:28 crc kubenswrapper[4809]: I1205 11:31:28.999803 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.106924 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2xsf\" (UniqueName: \"kubernetes.io/projected/a73984ba-8f10-4cc7-aa62-30fa504ab55f-kube-api-access-c2xsf\") pod \"nova-cell1-5fdb-account-create-update-92tdn\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.106991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a73984ba-8f10-4cc7-aa62-30fa504ab55f-operator-scripts\") pod \"nova-cell1-5fdb-account-create-update-92tdn\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.112283 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.120028 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.148817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerStarted","Data":"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61"} Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.149045 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerStarted","Data":"c9f0655ebdcf793c3e27fc709d02ef773ddb54e565e3b173984dbf181b003d08"} Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.208426 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2xsf\" (UniqueName: \"kubernetes.io/projected/a73984ba-8f10-4cc7-aa62-30fa504ab55f-kube-api-access-c2xsf\") pod \"nova-cell1-5fdb-account-create-update-92tdn\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.208474 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a73984ba-8f10-4cc7-aa62-30fa504ab55f-operator-scripts\") pod \"nova-cell1-5fdb-account-create-update-92tdn\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.209522 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a73984ba-8f10-4cc7-aa62-30fa504ab55f-operator-scripts\") pod \"nova-cell1-5fdb-account-create-update-92tdn\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.230393 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2xsf\" (UniqueName: \"kubernetes.io/projected/a73984ba-8f10-4cc7-aa62-30fa504ab55f-kube-api-access-c2xsf\") pod \"nova-cell1-5fdb-account-create-update-92tdn\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.259921 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.288787 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6kb6p"] Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.544394 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-h8l7j"] Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.705827 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ktb2f"] Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.954722 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-293e-account-create-update-82lc7"] Dec 05 11:31:29 crc kubenswrapper[4809]: I1205 11:31:29.991736 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0f47-account-create-update-f7kxx"] Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.073348 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5fdb-account-create-update-92tdn"] Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.213790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-h8l7j" event={"ID":"2b9c9e47-6fdc-4f5b-845f-503cf803036a","Type":"ContainerStarted","Data":"fd7d9a735a223752ce2da66b2a6e5377eebafb13bb894d79878a71ea6f44a4a7"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.214025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-h8l7j" event={"ID":"2b9c9e47-6fdc-4f5b-845f-503cf803036a","Type":"ContainerStarted","Data":"75bc70dc0916dd75c7556d4b8be6b93295250198908cd41f6c84d174763cb4e0"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.234922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-293e-account-create-update-82lc7" event={"ID":"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24","Type":"ContainerStarted","Data":"f5d263e99e7f315a6b5e6e91e0fb1dba79a81d9ac410379d23417b75fc916fb6"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.243366 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-h8l7j" podStartSLOduration=2.243346241 podStartE2EDuration="2.243346241s" podCreationTimestamp="2025-12-05 11:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:30.235032762 +0000 UTC m=+1385.626009320" watchObservedRunningTime="2025-12-05 11:31:30.243346241 +0000 UTC m=+1385.634322799" Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.247857 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" event={"ID":"b3931783-c543-4ec1-b91c-3b2a60c4b177","Type":"ContainerStarted","Data":"ae544935109404097875a98ff36988bb32e4ffe89e1e0ae10d538854a66ac2eb"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.259546 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6kb6p" event={"ID":"6c316081-5125-4ba2-8a5d-87174b3066d2","Type":"ContainerStarted","Data":"9a0cf769fac3a0d40431db27c3e061f9a49d341f34de1e9d46c4e22c316f1114"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.259590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6kb6p" event={"ID":"6c316081-5125-4ba2-8a5d-87174b3066d2","Type":"ContainerStarted","Data":"42d93e83b35f824e05ffaea32042f2e1b0555d9958fbf7ea64a83b42d6e1be4c"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.273271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerStarted","Data":"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.278585 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-6kb6p" podStartSLOduration=2.278568875 podStartE2EDuration="2.278568875s" podCreationTimestamp="2025-12-05 11:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:30.278045911 +0000 UTC m=+1385.669022479" watchObservedRunningTime="2025-12-05 11:31:30.278568875 +0000 UTC m=+1385.669545433" Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.283135 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" event={"ID":"a73984ba-8f10-4cc7-aa62-30fa504ab55f","Type":"ContainerStarted","Data":"f2de899760317ce859d078a84dfc9e9a3451e49563630b56e935877d0f3070bf"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.294380 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ktb2f" event={"ID":"f0813b74-3d43-48a9-a9af-738df2fe08ad","Type":"ContainerStarted","Data":"5c390c0b9a8e0fff7d4bd886bd35f57af0e1ddf10c0be644a90fe8d4d56cf5e7"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.300101 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21e68543-4956-4fc0-86aa-9ae8484b07c3","Type":"ContainerStarted","Data":"4770b253a022abba8402b06255199087a163db712ada152cea565a2565eb0ffc"} Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.300937 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.317728 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-ktb2f" podStartSLOduration=2.317706952 podStartE2EDuration="2.317706952s" podCreationTimestamp="2025-12-05 11:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:30.312214028 +0000 UTC m=+1385.703190596" watchObservedRunningTime="2025-12-05 11:31:30.317706952 +0000 UTC m=+1385.708683520" Dec 05 11:31:30 crc kubenswrapper[4809]: I1205 11:31:30.388333 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.645145563 podStartE2EDuration="3.388312064s" podCreationTimestamp="2025-12-05 11:31:27 +0000 UTC" firstStartedPulling="2025-12-05 11:31:28.059111145 +0000 UTC m=+1383.450087713" lastFinishedPulling="2025-12-05 11:31:28.802277656 +0000 UTC m=+1384.193254214" observedRunningTime="2025-12-05 11:31:30.328081884 +0000 UTC m=+1385.719058442" watchObservedRunningTime="2025-12-05 11:31:30.388312064 +0000 UTC m=+1385.779288622" Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.310921 4809 generic.go:334] "Generic (PLEG): container finished" podID="a73984ba-8f10-4cc7-aa62-30fa504ab55f" containerID="258c92cf388669a38f82f3d4a6a139c7c0d61b4d6ea7930faec41f22a47d2809" exitCode=0 Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.311129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" event={"ID":"a73984ba-8f10-4cc7-aa62-30fa504ab55f","Type":"ContainerDied","Data":"258c92cf388669a38f82f3d4a6a139c7c0d61b4d6ea7930faec41f22a47d2809"} Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.315737 4809 generic.go:334] "Generic (PLEG): container finished" podID="f0813b74-3d43-48a9-a9af-738df2fe08ad" containerID="2c109d5ee7420de221d8af2681db967c2c239aa7aa67e96c11e16f774976f28d" exitCode=0 Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.315885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ktb2f" event={"ID":"f0813b74-3d43-48a9-a9af-738df2fe08ad","Type":"ContainerDied","Data":"2c109d5ee7420de221d8af2681db967c2c239aa7aa67e96c11e16f774976f28d"} Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.321088 4809 generic.go:334] "Generic (PLEG): container finished" podID="2971236f-8f3a-4a90-bfcd-bfcdab8e1a24" containerID="6dc5d0d87115d603eb8dd2e73cec49003fe6ab7534b7c9f6a2b38814ea7d4fcb" exitCode=0 Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.321187 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-293e-account-create-update-82lc7" event={"ID":"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24","Type":"ContainerDied","Data":"6dc5d0d87115d603eb8dd2e73cec49003fe6ab7534b7c9f6a2b38814ea7d4fcb"} Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.327421 4809 generic.go:334] "Generic (PLEG): container finished" podID="b3931783-c543-4ec1-b91c-3b2a60c4b177" containerID="eddaf5222111460d42f2358e99f093a8dd969f342da92ec3db6de0ad73712f04" exitCode=0 Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.327618 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" event={"ID":"b3931783-c543-4ec1-b91c-3b2a60c4b177","Type":"ContainerDied","Data":"eddaf5222111460d42f2358e99f093a8dd969f342da92ec3db6de0ad73712f04"} Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.330026 4809 generic.go:334] "Generic (PLEG): container finished" podID="6c316081-5125-4ba2-8a5d-87174b3066d2" containerID="9a0cf769fac3a0d40431db27c3e061f9a49d341f34de1e9d46c4e22c316f1114" exitCode=0 Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.330077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6kb6p" event={"ID":"6c316081-5125-4ba2-8a5d-87174b3066d2","Type":"ContainerDied","Data":"9a0cf769fac3a0d40431db27c3e061f9a49d341f34de1e9d46c4e22c316f1114"} Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.335292 4809 generic.go:334] "Generic (PLEG): container finished" podID="2b9c9e47-6fdc-4f5b-845f-503cf803036a" containerID="fd7d9a735a223752ce2da66b2a6e5377eebafb13bb894d79878a71ea6f44a4a7" exitCode=0 Dec 05 11:31:31 crc kubenswrapper[4809]: I1205 11:31:31.335809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-h8l7j" event={"ID":"2b9c9e47-6fdc-4f5b-845f-503cf803036a","Type":"ContainerDied","Data":"fd7d9a735a223752ce2da66b2a6e5377eebafb13bb894d79878a71ea6f44a4a7"} Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.345435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerStarted","Data":"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d"} Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.753065 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.793284 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3931783-c543-4ec1-b91c-3b2a60c4b177-operator-scripts\") pod \"b3931783-c543-4ec1-b91c-3b2a60c4b177\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.793657 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6mrc\" (UniqueName: \"kubernetes.io/projected/b3931783-c543-4ec1-b91c-3b2a60c4b177-kube-api-access-f6mrc\") pod \"b3931783-c543-4ec1-b91c-3b2a60c4b177\" (UID: \"b3931783-c543-4ec1-b91c-3b2a60c4b177\") " Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.794084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3931783-c543-4ec1-b91c-3b2a60c4b177-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3931783-c543-4ec1-b91c-3b2a60c4b177" (UID: "b3931783-c543-4ec1-b91c-3b2a60c4b177"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.810930 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3931783-c543-4ec1-b91c-3b2a60c4b177-kube-api-access-f6mrc" (OuterVolumeSpecName: "kube-api-access-f6mrc") pod "b3931783-c543-4ec1-b91c-3b2a60c4b177" (UID: "b3931783-c543-4ec1-b91c-3b2a60c4b177"). InnerVolumeSpecName "kube-api-access-f6mrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.896358 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6mrc\" (UniqueName: \"kubernetes.io/projected/b3931783-c543-4ec1-b91c-3b2a60c4b177-kube-api-access-f6mrc\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.897074 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3931783-c543-4ec1-b91c-3b2a60c4b177-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.956264 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.964334 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.974448 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:32 crc kubenswrapper[4809]: I1205 11:31:32.981200 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.002992 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102214 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b9c9e47-6fdc-4f5b-845f-503cf803036a-operator-scripts\") pod \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102294 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c316081-5125-4ba2-8a5d-87174b3066d2-operator-scripts\") pod \"6c316081-5125-4ba2-8a5d-87174b3066d2\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102356 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhsks\" (UniqueName: \"kubernetes.io/projected/2b9c9e47-6fdc-4f5b-845f-503cf803036a-kube-api-access-xhsks\") pod \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\" (UID: \"2b9c9e47-6fdc-4f5b-845f-503cf803036a\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102379 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a73984ba-8f10-4cc7-aa62-30fa504ab55f-operator-scripts\") pod \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0813b74-3d43-48a9-a9af-738df2fe08ad-operator-scripts\") pod \"f0813b74-3d43-48a9-a9af-738df2fe08ad\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102537 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-operator-scripts\") pod \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102571 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rh2g\" (UniqueName: \"kubernetes.io/projected/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-kube-api-access-2rh2g\") pod \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\" (UID: \"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102666 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2xsf\" (UniqueName: \"kubernetes.io/projected/a73984ba-8f10-4cc7-aa62-30fa504ab55f-kube-api-access-c2xsf\") pod \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\" (UID: \"a73984ba-8f10-4cc7-aa62-30fa504ab55f\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102689 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkxs7\" (UniqueName: \"kubernetes.io/projected/6c316081-5125-4ba2-8a5d-87174b3066d2-kube-api-access-xkxs7\") pod \"6c316081-5125-4ba2-8a5d-87174b3066d2\" (UID: \"6c316081-5125-4ba2-8a5d-87174b3066d2\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.102720 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zxmt\" (UniqueName: \"kubernetes.io/projected/f0813b74-3d43-48a9-a9af-738df2fe08ad-kube-api-access-5zxmt\") pod \"f0813b74-3d43-48a9-a9af-738df2fe08ad\" (UID: \"f0813b74-3d43-48a9-a9af-738df2fe08ad\") " Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.104061 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0813b74-3d43-48a9-a9af-738df2fe08ad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0813b74-3d43-48a9-a9af-738df2fe08ad" (UID: "f0813b74-3d43-48a9-a9af-738df2fe08ad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.104361 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9c9e47-6fdc-4f5b-845f-503cf803036a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b9c9e47-6fdc-4f5b-845f-503cf803036a" (UID: "2b9c9e47-6fdc-4f5b-845f-503cf803036a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.104684 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c316081-5125-4ba2-8a5d-87174b3066d2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c316081-5125-4ba2-8a5d-87174b3066d2" (UID: "6c316081-5125-4ba2-8a5d-87174b3066d2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.110799 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73984ba-8f10-4cc7-aa62-30fa504ab55f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a73984ba-8f10-4cc7-aa62-30fa504ab55f" (UID: "a73984ba-8f10-4cc7-aa62-30fa504ab55f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.110967 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b9c9e47-6fdc-4f5b-845f-503cf803036a-kube-api-access-xhsks" (OuterVolumeSpecName: "kube-api-access-xhsks") pod "2b9c9e47-6fdc-4f5b-845f-503cf803036a" (UID: "2b9c9e47-6fdc-4f5b-845f-503cf803036a"). InnerVolumeSpecName "kube-api-access-xhsks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.111423 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2971236f-8f3a-4a90-bfcd-bfcdab8e1a24" (UID: "2971236f-8f3a-4a90-bfcd-bfcdab8e1a24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.112833 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-kube-api-access-2rh2g" (OuterVolumeSpecName: "kube-api-access-2rh2g") pod "2971236f-8f3a-4a90-bfcd-bfcdab8e1a24" (UID: "2971236f-8f3a-4a90-bfcd-bfcdab8e1a24"). InnerVolumeSpecName "kube-api-access-2rh2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.114577 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73984ba-8f10-4cc7-aa62-30fa504ab55f-kube-api-access-c2xsf" (OuterVolumeSpecName: "kube-api-access-c2xsf") pod "a73984ba-8f10-4cc7-aa62-30fa504ab55f" (UID: "a73984ba-8f10-4cc7-aa62-30fa504ab55f"). InnerVolumeSpecName "kube-api-access-c2xsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.126422 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c316081-5125-4ba2-8a5d-87174b3066d2-kube-api-access-xkxs7" (OuterVolumeSpecName: "kube-api-access-xkxs7") pod "6c316081-5125-4ba2-8a5d-87174b3066d2" (UID: "6c316081-5125-4ba2-8a5d-87174b3066d2"). InnerVolumeSpecName "kube-api-access-xkxs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.129805 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0813b74-3d43-48a9-a9af-738df2fe08ad-kube-api-access-5zxmt" (OuterVolumeSpecName: "kube-api-access-5zxmt") pod "f0813b74-3d43-48a9-a9af-738df2fe08ad" (UID: "f0813b74-3d43-48a9-a9af-738df2fe08ad"). InnerVolumeSpecName "kube-api-access-5zxmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204724 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2xsf\" (UniqueName: \"kubernetes.io/projected/a73984ba-8f10-4cc7-aa62-30fa504ab55f-kube-api-access-c2xsf\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204762 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkxs7\" (UniqueName: \"kubernetes.io/projected/6c316081-5125-4ba2-8a5d-87174b3066d2-kube-api-access-xkxs7\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204774 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zxmt\" (UniqueName: \"kubernetes.io/projected/f0813b74-3d43-48a9-a9af-738df2fe08ad-kube-api-access-5zxmt\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204787 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b9c9e47-6fdc-4f5b-845f-503cf803036a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204800 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c316081-5125-4ba2-8a5d-87174b3066d2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204812 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a73984ba-8f10-4cc7-aa62-30fa504ab55f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204823 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhsks\" (UniqueName: \"kubernetes.io/projected/2b9c9e47-6fdc-4f5b-845f-503cf803036a-kube-api-access-xhsks\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204834 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0813b74-3d43-48a9-a9af-738df2fe08ad-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204844 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.204856 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rh2g\" (UniqueName: \"kubernetes.io/projected/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24-kube-api-access-2rh2g\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.406853 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kb6p" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.407704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6kb6p" event={"ID":"6c316081-5125-4ba2-8a5d-87174b3066d2","Type":"ContainerDied","Data":"42d93e83b35f824e05ffaea32042f2e1b0555d9958fbf7ea64a83b42d6e1be4c"} Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.407754 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42d93e83b35f824e05ffaea32042f2e1b0555d9958fbf7ea64a83b42d6e1be4c" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.410298 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-h8l7j" event={"ID":"2b9c9e47-6fdc-4f5b-845f-503cf803036a","Type":"ContainerDied","Data":"75bc70dc0916dd75c7556d4b8be6b93295250198908cd41f6c84d174763cb4e0"} Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.420339 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75bc70dc0916dd75c7556d4b8be6b93295250198908cd41f6c84d174763cb4e0" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.420309 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h8l7j" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.425792 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.425794 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5fdb-account-create-update-92tdn" event={"ID":"a73984ba-8f10-4cc7-aa62-30fa504ab55f","Type":"ContainerDied","Data":"f2de899760317ce859d078a84dfc9e9a3451e49563630b56e935877d0f3070bf"} Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.425835 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2de899760317ce859d078a84dfc9e9a3451e49563630b56e935877d0f3070bf" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.428728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ktb2f" event={"ID":"f0813b74-3d43-48a9-a9af-738df2fe08ad","Type":"ContainerDied","Data":"5c390c0b9a8e0fff7d4bd886bd35f57af0e1ddf10c0be644a90fe8d4d56cf5e7"} Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.428774 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c390c0b9a8e0fff7d4bd886bd35f57af0e1ddf10c0be644a90fe8d4d56cf5e7" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.428872 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ktb2f" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.432565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-293e-account-create-update-82lc7" event={"ID":"2971236f-8f3a-4a90-bfcd-bfcdab8e1a24","Type":"ContainerDied","Data":"f5d263e99e7f315a6b5e6e91e0fb1dba79a81d9ac410379d23417b75fc916fb6"} Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.432612 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5d263e99e7f315a6b5e6e91e0fb1dba79a81d9ac410379d23417b75fc916fb6" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.432699 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-293e-account-create-update-82lc7" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.435310 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" event={"ID":"b3931783-c543-4ec1-b91c-3b2a60c4b177","Type":"ContainerDied","Data":"ae544935109404097875a98ff36988bb32e4ffe89e1e0ae10d538854a66ac2eb"} Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.435344 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae544935109404097875a98ff36988bb32e4ffe89e1e0ae10d538854a66ac2eb" Dec 05 11:31:33 crc kubenswrapper[4809]: I1205 11:31:33.436487 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0f47-account-create-update-f7kxx" Dec 05 11:31:34 crc kubenswrapper[4809]: I1205 11:31:34.448399 4809 generic.go:334] "Generic (PLEG): container finished" podID="50427365-4dce-49d9-8d35-09866162c6de" containerID="1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9" exitCode=1 Dec 05 11:31:34 crc kubenswrapper[4809]: I1205 11:31:34.448596 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-central-agent" containerID="cri-o://e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61" gracePeriod=30 Dec 05 11:31:34 crc kubenswrapper[4809]: I1205 11:31:34.448686 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-notification-agent" containerID="cri-o://ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c" gracePeriod=30 Dec 05 11:31:34 crc kubenswrapper[4809]: I1205 11:31:34.448515 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerDied","Data":"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9"} Dec 05 11:31:34 crc kubenswrapper[4809]: I1205 11:31:34.450513 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="sg-core" containerID="cri-o://9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d" gracePeriod=30 Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.227426 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.340711 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-log-httpd\") pod \"50427365-4dce-49d9-8d35-09866162c6de\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.340811 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-run-httpd\") pod \"50427365-4dce-49d9-8d35-09866162c6de\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.340875 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-scripts\") pod \"50427365-4dce-49d9-8d35-09866162c6de\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.340941 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-sg-core-conf-yaml\") pod \"50427365-4dce-49d9-8d35-09866162c6de\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.341057 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-config-data\") pod \"50427365-4dce-49d9-8d35-09866162c6de\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.341102 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-combined-ca-bundle\") pod \"50427365-4dce-49d9-8d35-09866162c6de\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.341138 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l5qq\" (UniqueName: \"kubernetes.io/projected/50427365-4dce-49d9-8d35-09866162c6de-kube-api-access-7l5qq\") pod \"50427365-4dce-49d9-8d35-09866162c6de\" (UID: \"50427365-4dce-49d9-8d35-09866162c6de\") " Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.341319 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "50427365-4dce-49d9-8d35-09866162c6de" (UID: "50427365-4dce-49d9-8d35-09866162c6de"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.341664 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.342285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "50427365-4dce-49d9-8d35-09866162c6de" (UID: "50427365-4dce-49d9-8d35-09866162c6de"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.346416 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50427365-4dce-49d9-8d35-09866162c6de-kube-api-access-7l5qq" (OuterVolumeSpecName: "kube-api-access-7l5qq") pod "50427365-4dce-49d9-8d35-09866162c6de" (UID: "50427365-4dce-49d9-8d35-09866162c6de"). InnerVolumeSpecName "kube-api-access-7l5qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.351713 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-scripts" (OuterVolumeSpecName: "scripts") pod "50427365-4dce-49d9-8d35-09866162c6de" (UID: "50427365-4dce-49d9-8d35-09866162c6de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.387736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "50427365-4dce-49d9-8d35-09866162c6de" (UID: "50427365-4dce-49d9-8d35-09866162c6de"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.402253 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.402526 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-log" containerID="cri-o://a7397be14fac2da1c585e4bf469b3f7e487a264784a94fc89ff5aa44fb1cf14a" gracePeriod=30 Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.403030 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-httpd" containerID="cri-o://fa1e16959dbb8870e1c65a26b85f2c8b1bd0fa38bd4406d6708a3c594ed2f0c5" gracePeriod=30 Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.443568 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l5qq\" (UniqueName: \"kubernetes.io/projected/50427365-4dce-49d9-8d35-09866162c6de-kube-api-access-7l5qq\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.443601 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50427365-4dce-49d9-8d35-09866162c6de-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.443614 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.443623 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.464873 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50427365-4dce-49d9-8d35-09866162c6de" (UID: "50427365-4dce-49d9-8d35-09866162c6de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.465920 4809 generic.go:334] "Generic (PLEG): container finished" podID="50427365-4dce-49d9-8d35-09866162c6de" containerID="9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d" exitCode=2 Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.465949 4809 generic.go:334] "Generic (PLEG): container finished" podID="50427365-4dce-49d9-8d35-09866162c6de" containerID="ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c" exitCode=0 Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.465962 4809 generic.go:334] "Generic (PLEG): container finished" podID="50427365-4dce-49d9-8d35-09866162c6de" containerID="e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61" exitCode=0 Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.465983 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerDied","Data":"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d"} Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.466014 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerDied","Data":"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c"} Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.466025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerDied","Data":"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61"} Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.466034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50427365-4dce-49d9-8d35-09866162c6de","Type":"ContainerDied","Data":"c9f0655ebdcf793c3e27fc709d02ef773ddb54e565e3b173984dbf181b003d08"} Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.466049 4809 scope.go:117] "RemoveContainer" containerID="1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.466171 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.497762 4809 scope.go:117] "RemoveContainer" containerID="9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.506734 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-config-data" (OuterVolumeSpecName: "config-data") pod "50427365-4dce-49d9-8d35-09866162c6de" (UID: "50427365-4dce-49d9-8d35-09866162c6de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.521373 4809 scope.go:117] "RemoveContainer" containerID="ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.544720 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.544744 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50427365-4dce-49d9-8d35-09866162c6de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.547999 4809 scope.go:117] "RemoveContainer" containerID="e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.566777 4809 scope.go:117] "RemoveContainer" containerID="1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.567248 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9\": container with ID starting with 1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9 not found: ID does not exist" containerID="1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.567287 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9"} err="failed to get container status \"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9\": rpc error: code = NotFound desc = could not find container \"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9\": container with ID starting with 1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9 not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.567312 4809 scope.go:117] "RemoveContainer" containerID="9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.567700 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d\": container with ID starting with 9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d not found: ID does not exist" containerID="9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.567720 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d"} err="failed to get container status \"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d\": rpc error: code = NotFound desc = could not find container \"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d\": container with ID starting with 9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.567734 4809 scope.go:117] "RemoveContainer" containerID="ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.567945 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c\": container with ID starting with ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c not found: ID does not exist" containerID="ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.567965 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c"} err="failed to get container status \"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c\": rpc error: code = NotFound desc = could not find container \"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c\": container with ID starting with ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.567977 4809 scope.go:117] "RemoveContainer" containerID="e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.568195 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61\": container with ID starting with e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61 not found: ID does not exist" containerID="e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.568213 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61"} err="failed to get container status \"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61\": rpc error: code = NotFound desc = could not find container \"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61\": container with ID starting with e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61 not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.568225 4809 scope.go:117] "RemoveContainer" containerID="1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.568764 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9"} err="failed to get container status \"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9\": rpc error: code = NotFound desc = could not find container \"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9\": container with ID starting with 1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9 not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.568816 4809 scope.go:117] "RemoveContainer" containerID="9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569165 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d"} err="failed to get container status \"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d\": rpc error: code = NotFound desc = could not find container \"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d\": container with ID starting with 9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569185 4809 scope.go:117] "RemoveContainer" containerID="ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569400 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c"} err="failed to get container status \"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c\": rpc error: code = NotFound desc = could not find container \"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c\": container with ID starting with ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569413 4809 scope.go:117] "RemoveContainer" containerID="e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569702 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61"} err="failed to get container status \"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61\": rpc error: code = NotFound desc = could not find container \"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61\": container with ID starting with e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61 not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569722 4809 scope.go:117] "RemoveContainer" containerID="1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569962 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9"} err="failed to get container status \"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9\": rpc error: code = NotFound desc = could not find container \"1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9\": container with ID starting with 1cc5d70be0c58252a0577b70d33c95994d935b7aef768429f57a4894313b00d9 not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.569994 4809 scope.go:117] "RemoveContainer" containerID="9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.570271 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d"} err="failed to get container status \"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d\": rpc error: code = NotFound desc = could not find container \"9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d\": container with ID starting with 9a17ef0fc6f4b5fe3327a654c30192a5237772eab7f16b78858d7dd72ef2d36d not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.570309 4809 scope.go:117] "RemoveContainer" containerID="ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.570968 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c"} err="failed to get container status \"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c\": rpc error: code = NotFound desc = could not find container \"ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c\": container with ID starting with ff7e2ca8c9f8dc6c70b592b7b92bf7f7117e26da575997042c5754176a0d1a6c not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.571019 4809 scope.go:117] "RemoveContainer" containerID="e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.571321 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61"} err="failed to get container status \"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61\": rpc error: code = NotFound desc = could not find container \"e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61\": container with ID starting with e733cff7ff5b85673ecf28db80a202150d363e62c96d1dce2bcdd4948f4eeb61 not found: ID does not exist" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.830768 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.837881 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.862827 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.863585 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c316081-5125-4ba2-8a5d-87174b3066d2" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.863608 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c316081-5125-4ba2-8a5d-87174b3066d2" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.863838 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3931783-c543-4ec1-b91c-3b2a60c4b177" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.863850 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3931783-c543-4ec1-b91c-3b2a60c4b177" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.863884 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="proxy-httpd" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.863892 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="proxy-httpd" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.863926 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2971236f-8f3a-4a90-bfcd-bfcdab8e1a24" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.863934 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2971236f-8f3a-4a90-bfcd-bfcdab8e1a24" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.863943 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-central-agent" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.863952 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-central-agent" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.863975 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-notification-agent" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.863985 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-notification-agent" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.863995 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="sg-core" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864002 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="sg-core" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.864020 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0813b74-3d43-48a9-a9af-738df2fe08ad" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864027 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0813b74-3d43-48a9-a9af-738df2fe08ad" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.864063 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b9c9e47-6fdc-4f5b-845f-503cf803036a" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864071 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b9c9e47-6fdc-4f5b-845f-503cf803036a" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: E1205 11:31:35.864095 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73984ba-8f10-4cc7-aa62-30fa504ab55f" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864103 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73984ba-8f10-4cc7-aa62-30fa504ab55f" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864538 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="proxy-httpd" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864561 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2971236f-8f3a-4a90-bfcd-bfcdab8e1a24" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864598 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c316081-5125-4ba2-8a5d-87174b3066d2" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864649 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3931783-c543-4ec1-b91c-3b2a60c4b177" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864677 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-central-agent" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864696 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="ceilometer-notification-agent" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864730 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73984ba-8f10-4cc7-aa62-30fa504ab55f" containerName="mariadb-account-create-update" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864747 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50427365-4dce-49d9-8d35-09866162c6de" containerName="sg-core" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864775 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0813b74-3d43-48a9-a9af-738df2fe08ad" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.864802 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b9c9e47-6fdc-4f5b-845f-503cf803036a" containerName="mariadb-database-create" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.869149 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.878954 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.879095 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.879250 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.928172 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.956887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.957021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-run-httpd\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.957060 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-log-httpd\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.957092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.957120 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.957147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhlcm\" (UniqueName: \"kubernetes.io/projected/50b443ed-955a-40d1-9aa1-ab0036437131-kube-api-access-qhlcm\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.957212 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-config-data\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:35 crc kubenswrapper[4809]: I1205 11:31:35.957251 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-scripts\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.058716 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.058770 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.058794 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhlcm\" (UniqueName: \"kubernetes.io/projected/50b443ed-955a-40d1-9aa1-ab0036437131-kube-api-access-qhlcm\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.058838 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-config-data\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.058871 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-scripts\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.058902 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.058979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-run-httpd\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.059012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-log-httpd\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.059480 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-log-httpd\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.067002 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-run-httpd\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.067080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.067595 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.074619 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-scripts\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.079375 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.089228 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-config-data\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.091684 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhlcm\" (UniqueName: \"kubernetes.io/projected/50b443ed-955a-40d1-9aa1-ab0036437131-kube-api-access-qhlcm\") pod \"ceilometer-0\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.195795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.478739 4809 generic.go:334] "Generic (PLEG): container finished" podID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerID="a7397be14fac2da1c585e4bf469b3f7e487a264784a94fc89ff5aa44fb1cf14a" exitCode=143 Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.478832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"157daa4d-2576-408d-a86d-c9c6e9c162e1","Type":"ContainerDied","Data":"a7397be14fac2da1c585e4bf469b3f7e487a264784a94fc89ff5aa44fb1cf14a"} Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.631418 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:36 crc kubenswrapper[4809]: W1205 11:31:36.678203 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50b443ed_955a_40d1_9aa1_ab0036437131.slice/crio-5851fe2a3e4cc155d54b053e23c1a438d373219d326f98fcf418f984b539d7bb WatchSource:0}: Error finding container 5851fe2a3e4cc155d54b053e23c1a438d373219d326f98fcf418f984b539d7bb: Status 404 returned error can't find the container with id 5851fe2a3e4cc155d54b053e23c1a438d373219d326f98fcf418f984b539d7bb Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.679075 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:36 crc kubenswrapper[4809]: I1205 11:31:36.882882 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50427365-4dce-49d9-8d35-09866162c6de" path="/var/lib/kubelet/pods/50427365-4dce-49d9-8d35-09866162c6de/volumes" Dec 05 11:31:37 crc kubenswrapper[4809]: I1205 11:31:37.489143 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerStarted","Data":"02f907fb4b72156322548335a3eb96075c6889eea7bc55abf6f73c4cc140df0c"} Dec 05 11:31:37 crc kubenswrapper[4809]: I1205 11:31:37.489460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerStarted","Data":"5851fe2a3e4cc155d54b053e23c1a438d373219d326f98fcf418f984b539d7bb"} Dec 05 11:31:37 crc kubenswrapper[4809]: I1205 11:31:37.609582 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 11:31:37 crc kubenswrapper[4809]: I1205 11:31:37.836972 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:31:37 crc kubenswrapper[4809]: I1205 11:31:37.837538 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-log" containerID="cri-o://596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99" gracePeriod=30 Dec 05 11:31:37 crc kubenswrapper[4809]: I1205 11:31:37.839117 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-httpd" containerID="cri-o://ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035" gracePeriod=30 Dec 05 11:31:38 crc kubenswrapper[4809]: I1205 11:31:38.507806 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerID="596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99" exitCode=143 Dec 05 11:31:38 crc kubenswrapper[4809]: I1205 11:31:38.508000 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2c4570ff-b46c-48e1-9aff-1b460ba27581","Type":"ContainerDied","Data":"596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99"} Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.354595 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2pxck"] Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.355847 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.357510 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.359515 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fc49v" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.359889 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.364848 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2pxck"] Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.527417 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.527482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcz6q\" (UniqueName: \"kubernetes.io/projected/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-kube-api-access-xcz6q\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.527737 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-scripts\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.527923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-config-data\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.633023 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-scripts\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.633194 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-config-data\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.633372 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.633472 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcz6q\" (UniqueName: \"kubernetes.io/projected/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-kube-api-access-xcz6q\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.640920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-scripts\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.642991 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.648911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-config-data\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.664646 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcz6q\" (UniqueName: \"kubernetes.io/projected/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-kube-api-access-xcz6q\") pod \"nova-cell0-conductor-db-sync-2pxck\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:39 crc kubenswrapper[4809]: I1205 11:31:39.672746 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:31:40 crc kubenswrapper[4809]: I1205 11:31:40.130218 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2pxck"] Dec 05 11:31:40 crc kubenswrapper[4809]: W1205 11:31:40.181581 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aa2367d_dad7_461e_9e93_1ebc2bfee6e8.slice/crio-6305823f18f4277ffc6bc86133f2d62ee65779ea4a52d25d602c9e38aaa1b86c WatchSource:0}: Error finding container 6305823f18f4277ffc6bc86133f2d62ee65779ea4a52d25d602c9e38aaa1b86c: Status 404 returned error can't find the container with id 6305823f18f4277ffc6bc86133f2d62ee65779ea4a52d25d602c9e38aaa1b86c Dec 05 11:31:40 crc kubenswrapper[4809]: I1205 11:31:40.530894 4809 generic.go:334] "Generic (PLEG): container finished" podID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerID="fa1e16959dbb8870e1c65a26b85f2c8b1bd0fa38bd4406d6708a3c594ed2f0c5" exitCode=0 Dec 05 11:31:40 crc kubenswrapper[4809]: I1205 11:31:40.530985 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"157daa4d-2576-408d-a86d-c9c6e9c162e1","Type":"ContainerDied","Data":"fa1e16959dbb8870e1c65a26b85f2c8b1bd0fa38bd4406d6708a3c594ed2f0c5"} Dec 05 11:31:40 crc kubenswrapper[4809]: I1205 11:31:40.534530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerStarted","Data":"84d04f0c8013a1b34d97494a4baab13b497cc65dd80418fdba85c7650521c422"} Dec 05 11:31:40 crc kubenswrapper[4809]: I1205 11:31:40.536122 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2pxck" event={"ID":"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8","Type":"ContainerStarted","Data":"6305823f18f4277ffc6bc86133f2d62ee65779ea4a52d25d602c9e38aaa1b86c"} Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.002807 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165049 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-scripts\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165093 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-logs\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165127 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-config-data\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165153 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165188 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-public-tls-certs\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165219 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-httpd-run\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165303 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-combined-ca-bundle\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165334 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58s8m\" (UniqueName: \"kubernetes.io/projected/157daa4d-2576-408d-a86d-c9c6e9c162e1-kube-api-access-58s8m\") pod \"157daa4d-2576-408d-a86d-c9c6e9c162e1\" (UID: \"157daa4d-2576-408d-a86d-c9c6e9c162e1\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.165845 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-logs" (OuterVolumeSpecName: "logs") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.166078 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.171446 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.172756 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-scripts" (OuterVolumeSpecName: "scripts") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.191432 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/157daa4d-2576-408d-a86d-c9c6e9c162e1-kube-api-access-58s8m" (OuterVolumeSpecName: "kube-api-access-58s8m") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "kube-api-access-58s8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.232398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.267300 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.267333 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58s8m\" (UniqueName: \"kubernetes.io/projected/157daa4d-2576-408d-a86d-c9c6e9c162e1-kube-api-access-58s8m\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.267346 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.267355 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.267376 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.267385 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/157daa4d-2576-408d-a86d-c9c6e9c162e1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.285428 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.307993 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.315736 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-config-data" (OuterVolumeSpecName: "config-data") pod "157daa4d-2576-408d-a86d-c9c6e9c162e1" (UID: "157daa4d-2576-408d-a86d-c9c6e9c162e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.371048 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.371087 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.371099 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157daa4d-2576-408d-a86d-c9c6e9c162e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.538539 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.553890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"157daa4d-2576-408d-a86d-c9c6e9c162e1","Type":"ContainerDied","Data":"68611642e279ddb57dc391443fe6cccefdb2744dd861c8937d91fb5911a8eaaf"} Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.554228 4809 scope.go:117] "RemoveContainer" containerID="fa1e16959dbb8870e1c65a26b85f2c8b1bd0fa38bd4406d6708a3c594ed2f0c5" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.554368 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.557132 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerStarted","Data":"8de9dfe27cdf5d1255a822d29bf9a8baa35416ab7b2f6225f760f23cd61d1100"} Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.565129 4809 generic.go:334] "Generic (PLEG): container finished" podID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerID="ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035" exitCode=0 Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.565196 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2c4570ff-b46c-48e1-9aff-1b460ba27581","Type":"ContainerDied","Data":"ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035"} Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.565233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2c4570ff-b46c-48e1-9aff-1b460ba27581","Type":"ContainerDied","Data":"832e6b0b010a706b9054e10290dfc7e15e925ecc0a20eb3c6ca0ed3c872b50ef"} Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.565346 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.608580 4809 scope.go:117] "RemoveContainer" containerID="a7397be14fac2da1c585e4bf469b3f7e487a264784a94fc89ff5aa44fb1cf14a" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.608735 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.624839 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.635985 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:31:41 crc kubenswrapper[4809]: E1205 11:31:41.636400 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-log" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636412 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-log" Dec 05 11:31:41 crc kubenswrapper[4809]: E1205 11:31:41.636438 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-httpd" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636445 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-httpd" Dec 05 11:31:41 crc kubenswrapper[4809]: E1205 11:31:41.636459 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-httpd" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636465 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-httpd" Dec 05 11:31:41 crc kubenswrapper[4809]: E1205 11:31:41.636481 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-log" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636486 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-log" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636660 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-log" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636670 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-httpd" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636683 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" containerName="glance-httpd" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.636703 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" containerName="glance-log" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.638479 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.643278 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.643677 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.652283 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.679139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-scripts\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.679219 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-combined-ca-bundle\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.679301 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdx96\" (UniqueName: \"kubernetes.io/projected/2c4570ff-b46c-48e1-9aff-1b460ba27581-kube-api-access-tdx96\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.679348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-config-data\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.679424 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-logs\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.679447 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.683490 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-internal-tls-certs\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.683560 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-httpd-run\") pod \"2c4570ff-b46c-48e1-9aff-1b460ba27581\" (UID: \"2c4570ff-b46c-48e1-9aff-1b460ba27581\") " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.686515 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.686770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-logs" (OuterVolumeSpecName: "logs") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.687027 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-scripts" (OuterVolumeSpecName: "scripts") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.689875 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c4570ff-b46c-48e1-9aff-1b460ba27581-kube-api-access-tdx96" (OuterVolumeSpecName: "kube-api-access-tdx96") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "kube-api-access-tdx96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.690074 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.690729 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdx96\" (UniqueName: \"kubernetes.io/projected/2c4570ff-b46c-48e1-9aff-1b460ba27581-kube-api-access-tdx96\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.690760 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.690813 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.690824 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c4570ff-b46c-48e1-9aff-1b460ba27581-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.690833 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.728892 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.742619 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.752983 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.770324 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-config-data" (OuterVolumeSpecName: "config-data") pod "2c4570ff-b46c-48e1-9aff-1b460ba27581" (UID: "2c4570ff-b46c-48e1-9aff-1b460ba27581"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802483 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802616 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-scripts\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802654 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802685 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-logs\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802744 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-config-data\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802793 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5x8m\" (UniqueName: \"kubernetes.io/projected/cf8da287-1c2a-4d39-9d1b-dc5218b44907-kube-api-access-w5x8m\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802861 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802875 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802888 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.802900 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c4570ff-b46c-48e1-9aff-1b460ba27581-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.873847 4809 scope.go:117] "RemoveContainer" containerID="ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.907864 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.907923 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.907956 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-scripts\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.907973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.908000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.908022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-logs\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.908048 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-config-data\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.908082 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5x8m\" (UniqueName: \"kubernetes.io/projected/cf8da287-1c2a-4d39-9d1b-dc5218b44907-kube-api-access-w5x8m\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.915561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.915856 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.915856 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.918139 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-logs\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.924897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-config-data\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.938426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.954155 4809 scope.go:117] "RemoveContainer" containerID="596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.960579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-scripts\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:41 crc kubenswrapper[4809]: I1205 11:31:41.963430 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5x8m\" (UniqueName: \"kubernetes.io/projected/cf8da287-1c2a-4d39-9d1b-dc5218b44907-kube-api-access-w5x8m\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.073926 4809 scope.go:117] "RemoveContainer" containerID="ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.074325 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.084003 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:31:42 crc kubenswrapper[4809]: E1205 11:31:42.092902 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035\": container with ID starting with ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035 not found: ID does not exist" containerID="ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.092996 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035"} err="failed to get container status \"ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035\": rpc error: code = NotFound desc = could not find container \"ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035\": container with ID starting with ed5267f9dc3b18a477821d6b326de37a83376099357cbd6a8bda8a270fb73035 not found: ID does not exist" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.093045 4809 scope.go:117] "RemoveContainer" containerID="596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99" Dec 05 11:31:42 crc kubenswrapper[4809]: E1205 11:31:42.094071 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99\": container with ID starting with 596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99 not found: ID does not exist" containerID="596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.094144 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99"} err="failed to get container status \"596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99\": rpc error: code = NotFound desc = could not find container \"596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99\": container with ID starting with 596f3204f8b7c64be37047a0b7dcbe605f5bc0e6754be7abecdd49f8be944d99 not found: ID does not exist" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.125246 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.128419 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.131398 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.131737 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.132770 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.139303 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " pod="openstack/glance-default-external-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.246555 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.247161 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhfzb\" (UniqueName: \"kubernetes.io/projected/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-kube-api-access-lhfzb\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.247213 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-logs\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.247250 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.247272 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.247299 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.247366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.247387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348721 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhfzb\" (UniqueName: \"kubernetes.io/projected/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-kube-api-access-lhfzb\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348777 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-logs\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348803 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348821 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.348944 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.350579 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-logs\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.350845 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.352467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.355969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.357783 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.358208 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.367686 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.376338 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhfzb\" (UniqueName: \"kubernetes.io/projected/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-kube-api-access-lhfzb\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.413199 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.441006 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.476927 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.603549 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerStarted","Data":"6b497c4c92354cc31f0c88f8c00a1c774df18fe1677fcf6162021b788e0927f6"} Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.603767 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-central-agent" containerID="cri-o://02f907fb4b72156322548335a3eb96075c6889eea7bc55abf6f73c4cc140df0c" gracePeriod=30 Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.603929 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.603987 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="proxy-httpd" containerID="cri-o://6b497c4c92354cc31f0c88f8c00a1c774df18fe1677fcf6162021b788e0927f6" gracePeriod=30 Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.604085 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="sg-core" containerID="cri-o://8de9dfe27cdf5d1255a822d29bf9a8baa35416ab7b2f6225f760f23cd61d1100" gracePeriod=30 Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.604115 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-notification-agent" containerID="cri-o://84d04f0c8013a1b34d97494a4baab13b497cc65dd80418fdba85c7650521c422" gracePeriod=30 Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.651427 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.19928868 podStartE2EDuration="7.651408139s" podCreationTimestamp="2025-12-05 11:31:35 +0000 UTC" firstStartedPulling="2025-12-05 11:31:36.681199565 +0000 UTC m=+1392.072176123" lastFinishedPulling="2025-12-05 11:31:42.133319024 +0000 UTC m=+1397.524295582" observedRunningTime="2025-12-05 11:31:42.621476384 +0000 UTC m=+1398.012452932" watchObservedRunningTime="2025-12-05 11:31:42.651408139 +0000 UTC m=+1398.042384687" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.886626 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="157daa4d-2576-408d-a86d-c9c6e9c162e1" path="/var/lib/kubelet/pods/157daa4d-2576-408d-a86d-c9c6e9c162e1/volumes" Dec 05 11:31:42 crc kubenswrapper[4809]: I1205 11:31:42.887504 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c4570ff-b46c-48e1-9aff-1b460ba27581" path="/var/lib/kubelet/pods/2c4570ff-b46c-48e1-9aff-1b460ba27581/volumes" Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.073154 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.579674 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:31:43 crc kubenswrapper[4809]: W1205 11:31:43.581361 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bf5e0d0_86bc_4728_8ad4_678b4b28e190.slice/crio-f0c1ffac9960b682a69cfc960ac40ad765fc8c81b9e4570edabb87578d4e2af5 WatchSource:0}: Error finding container f0c1ffac9960b682a69cfc960ac40ad765fc8c81b9e4570edabb87578d4e2af5: Status 404 returned error can't find the container with id f0c1ffac9960b682a69cfc960ac40ad765fc8c81b9e4570edabb87578d4e2af5 Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.616544 4809 generic.go:334] "Generic (PLEG): container finished" podID="50b443ed-955a-40d1-9aa1-ab0036437131" containerID="6b497c4c92354cc31f0c88f8c00a1c774df18fe1677fcf6162021b788e0927f6" exitCode=0 Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.616581 4809 generic.go:334] "Generic (PLEG): container finished" podID="50b443ed-955a-40d1-9aa1-ab0036437131" containerID="8de9dfe27cdf5d1255a822d29bf9a8baa35416ab7b2f6225f760f23cd61d1100" exitCode=2 Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.616593 4809 generic.go:334] "Generic (PLEG): container finished" podID="50b443ed-955a-40d1-9aa1-ab0036437131" containerID="84d04f0c8013a1b34d97494a4baab13b497cc65dd80418fdba85c7650521c422" exitCode=0 Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.616674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerDied","Data":"6b497c4c92354cc31f0c88f8c00a1c774df18fe1677fcf6162021b788e0927f6"} Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.616704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerDied","Data":"8de9dfe27cdf5d1255a822d29bf9a8baa35416ab7b2f6225f760f23cd61d1100"} Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.616715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerDied","Data":"84d04f0c8013a1b34d97494a4baab13b497cc65dd80418fdba85c7650521c422"} Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.620497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8bf5e0d0-86bc-4728-8ad4-678b4b28e190","Type":"ContainerStarted","Data":"f0c1ffac9960b682a69cfc960ac40ad765fc8c81b9e4570edabb87578d4e2af5"} Dec 05 11:31:43 crc kubenswrapper[4809]: I1205 11:31:43.624396 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf8da287-1c2a-4d39-9d1b-dc5218b44907","Type":"ContainerStarted","Data":"9c3771a79faa4fe599e47a439b6e595566c3c0558d1e7a143e2b85ca278d1f06"} Dec 05 11:31:44 crc kubenswrapper[4809]: I1205 11:31:44.648601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8bf5e0d0-86bc-4728-8ad4-678b4b28e190","Type":"ContainerStarted","Data":"aa994132555d0692b38e19880ffec4d3ede1737f451a8f05d6f024366f509dfe"} Dec 05 11:31:44 crc kubenswrapper[4809]: I1205 11:31:44.650842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf8da287-1c2a-4d39-9d1b-dc5218b44907","Type":"ContainerStarted","Data":"635c738f9169739928f1ce8f18732784dac36a5c39b6e36889932cf1eda8f9e3"} Dec 05 11:31:44 crc kubenswrapper[4809]: I1205 11:31:44.650908 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf8da287-1c2a-4d39-9d1b-dc5218b44907","Type":"ContainerStarted","Data":"35013bebb85eaa5b886089268483165db1836de375fb81c222beeadbb2a74fed"} Dec 05 11:31:44 crc kubenswrapper[4809]: I1205 11:31:44.677519 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.677498915 podStartE2EDuration="3.677498915s" podCreationTimestamp="2025-12-05 11:31:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:44.677451084 +0000 UTC m=+1400.068427642" watchObservedRunningTime="2025-12-05 11:31:44.677498915 +0000 UTC m=+1400.068475473" Dec 05 11:31:45 crc kubenswrapper[4809]: I1205 11:31:45.667781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8bf5e0d0-86bc-4728-8ad4-678b4b28e190","Type":"ContainerStarted","Data":"a82c49db9f73a1cf818cf2522001c715c552fa25f9dd1ad00a5785861ff76f4c"} Dec 05 11:31:45 crc kubenswrapper[4809]: I1205 11:31:45.687515 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.687500259 podStartE2EDuration="3.687500259s" podCreationTimestamp="2025-12-05 11:31:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:31:45.687042897 +0000 UTC m=+1401.078019455" watchObservedRunningTime="2025-12-05 11:31:45.687500259 +0000 UTC m=+1401.078476817" Dec 05 11:31:49 crc kubenswrapper[4809]: I1205 11:31:49.731039 4809 generic.go:334] "Generic (PLEG): container finished" podID="50b443ed-955a-40d1-9aa1-ab0036437131" containerID="02f907fb4b72156322548335a3eb96075c6889eea7bc55abf6f73c4cc140df0c" exitCode=0 Dec 05 11:31:49 crc kubenswrapper[4809]: I1205 11:31:49.731129 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerDied","Data":"02f907fb4b72156322548335a3eb96075c6889eea7bc55abf6f73c4cc140df0c"} Dec 05 11:31:50 crc kubenswrapper[4809]: I1205 11:31:50.956735 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040545 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-sg-core-conf-yaml\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040606 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-log-httpd\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040743 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-run-httpd\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040806 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhlcm\" (UniqueName: \"kubernetes.io/projected/50b443ed-955a-40d1-9aa1-ab0036437131-kube-api-access-qhlcm\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040841 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-scripts\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-config-data\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040903 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-ceilometer-tls-certs\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.040957 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-combined-ca-bundle\") pod \"50b443ed-955a-40d1-9aa1-ab0036437131\" (UID: \"50b443ed-955a-40d1-9aa1-ab0036437131\") " Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.043663 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.045955 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.053044 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b443ed-955a-40d1-9aa1-ab0036437131-kube-api-access-qhlcm" (OuterVolumeSpecName: "kube-api-access-qhlcm") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "kube-api-access-qhlcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.060880 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-scripts" (OuterVolumeSpecName: "scripts") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.097569 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.101753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.144167 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.144457 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50b443ed-955a-40d1-9aa1-ab0036437131-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.144471 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhlcm\" (UniqueName: \"kubernetes.io/projected/50b443ed-955a-40d1-9aa1-ab0036437131-kube-api-access-qhlcm\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.144484 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.144495 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.144503 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.155010 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.162916 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-config-data" (OuterVolumeSpecName: "config-data") pod "50b443ed-955a-40d1-9aa1-ab0036437131" (UID: "50b443ed-955a-40d1-9aa1-ab0036437131"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.246421 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.246462 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b443ed-955a-40d1-9aa1-ab0036437131-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.751926 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50b443ed-955a-40d1-9aa1-ab0036437131","Type":"ContainerDied","Data":"5851fe2a3e4cc155d54b053e23c1a438d373219d326f98fcf418f984b539d7bb"} Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.751983 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.751995 4809 scope.go:117] "RemoveContainer" containerID="6b497c4c92354cc31f0c88f8c00a1c774df18fe1677fcf6162021b788e0927f6" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.755439 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2pxck" event={"ID":"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8","Type":"ContainerStarted","Data":"227b5a95db28ac44992a513775b179933f3871062f8c30164d38b84a81d68994"} Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.799083 4809 scope.go:117] "RemoveContainer" containerID="8de9dfe27cdf5d1255a822d29bf9a8baa35416ab7b2f6225f760f23cd61d1100" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.814379 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-2pxck" podStartSLOduration=2.277944887 podStartE2EDuration="12.814344592s" podCreationTimestamp="2025-12-05 11:31:39 +0000 UTC" firstStartedPulling="2025-12-05 11:31:40.183714424 +0000 UTC m=+1395.574690982" lastFinishedPulling="2025-12-05 11:31:50.720114129 +0000 UTC m=+1406.111090687" observedRunningTime="2025-12-05 11:31:51.800703494 +0000 UTC m=+1407.191680052" watchObservedRunningTime="2025-12-05 11:31:51.814344592 +0000 UTC m=+1407.205321190" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.839527 4809 scope.go:117] "RemoveContainer" containerID="84d04f0c8013a1b34d97494a4baab13b497cc65dd80418fdba85c7650521c422" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.844623 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.863863 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.874534 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.874940 4809 scope.go:117] "RemoveContainer" containerID="02f907fb4b72156322548335a3eb96075c6889eea7bc55abf6f73c4cc140df0c" Dec 05 11:31:51 crc kubenswrapper[4809]: E1205 11:31:51.875027 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-central-agent" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875044 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-central-agent" Dec 05 11:31:51 crc kubenswrapper[4809]: E1205 11:31:51.875068 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="proxy-httpd" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875077 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="proxy-httpd" Dec 05 11:31:51 crc kubenswrapper[4809]: E1205 11:31:51.875094 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-notification-agent" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875103 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-notification-agent" Dec 05 11:31:51 crc kubenswrapper[4809]: E1205 11:31:51.875138 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="sg-core" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875146 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="sg-core" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875351 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-central-agent" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875373 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="sg-core" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875392 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="ceilometer-notification-agent" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.875411 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" containerName="proxy-httpd" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.877451 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.882982 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.883164 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.883278 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.885789 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-log-httpd\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958463 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-scripts\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-run-httpd\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958539 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958606 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958679 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-config-data\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958898 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:51 crc kubenswrapper[4809]: I1205 11:31:51.958944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqkj4\" (UniqueName: \"kubernetes.io/projected/139b3658-4df4-4cff-9a73-1e20624b8d8f-kube-api-access-dqkj4\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-log-httpd\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060772 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-scripts\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-run-httpd\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060874 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-config-data\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.060980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqkj4\" (UniqueName: \"kubernetes.io/projected/139b3658-4df4-4cff-9a73-1e20624b8d8f-kube-api-access-dqkj4\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.062272 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-run-httpd\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.063117 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-log-httpd\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.066428 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.066434 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.067466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-config-data\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.067576 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-scripts\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.077541 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.090216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqkj4\" (UniqueName: \"kubernetes.io/projected/139b3658-4df4-4cff-9a73-1e20624b8d8f-kube-api-access-dqkj4\") pod \"ceilometer-0\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.198281 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.442112 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.442440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.477546 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.477609 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.489557 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.516397 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.518391 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.522904 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.722282 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.767400 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerStarted","Data":"9d2e380550da32aeeabc3b3c3426824248fded444568d9aa18705297e7481a6a"} Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.767616 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.767657 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.767958 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.768191 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 11:31:52 crc kubenswrapper[4809]: I1205 11:31:52.885155 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b443ed-955a-40d1-9aa1-ab0036437131" path="/var/lib/kubelet/pods/50b443ed-955a-40d1-9aa1-ab0036437131/volumes" Dec 05 11:31:53 crc kubenswrapper[4809]: I1205 11:31:53.779208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerStarted","Data":"7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e"} Dec 05 11:31:54 crc kubenswrapper[4809]: I1205 11:31:54.790359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerStarted","Data":"c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a"} Dec 05 11:31:54 crc kubenswrapper[4809]: I1205 11:31:54.791034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerStarted","Data":"8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c"} Dec 05 11:31:54 crc kubenswrapper[4809]: I1205 11:31:54.935516 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 11:31:54 crc kubenswrapper[4809]: I1205 11:31:54.935622 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:31:55 crc kubenswrapper[4809]: I1205 11:31:55.019985 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:55 crc kubenswrapper[4809]: I1205 11:31:55.020068 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:31:55 crc kubenswrapper[4809]: I1205 11:31:55.024410 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 11:31:55 crc kubenswrapper[4809]: I1205 11:31:55.057379 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 11:31:56 crc kubenswrapper[4809]: I1205 11:31:56.832204 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerStarted","Data":"3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9"} Dec 05 11:31:56 crc kubenswrapper[4809]: I1205 11:31:56.833974 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 11:31:56 crc kubenswrapper[4809]: I1205 11:31:56.858579 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.676581079 podStartE2EDuration="5.858562777s" podCreationTimestamp="2025-12-05 11:31:51 +0000 UTC" firstStartedPulling="2025-12-05 11:31:52.730164214 +0000 UTC m=+1408.121140772" lastFinishedPulling="2025-12-05 11:31:55.912145912 +0000 UTC m=+1411.303122470" observedRunningTime="2025-12-05 11:31:56.854609713 +0000 UTC m=+1412.245586271" watchObservedRunningTime="2025-12-05 11:31:56.858562777 +0000 UTC m=+1412.249539335" Dec 05 11:32:03 crc kubenswrapper[4809]: I1205 11:32:03.898424 4809 generic.go:334] "Generic (PLEG): container finished" podID="4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" containerID="227b5a95db28ac44992a513775b179933f3871062f8c30164d38b84a81d68994" exitCode=0 Dec 05 11:32:03 crc kubenswrapper[4809]: I1205 11:32:03.898523 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2pxck" event={"ID":"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8","Type":"ContainerDied","Data":"227b5a95db28ac44992a513775b179933f3871062f8c30164d38b84a81d68994"} Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.388178 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.422119 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcz6q\" (UniqueName: \"kubernetes.io/projected/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-kube-api-access-xcz6q\") pod \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.422195 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-scripts\") pod \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.422247 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-config-data\") pod \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.422302 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-combined-ca-bundle\") pod \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\" (UID: \"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8\") " Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.428247 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-scripts" (OuterVolumeSpecName: "scripts") pod "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" (UID: "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.435840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-kube-api-access-xcz6q" (OuterVolumeSpecName: "kube-api-access-xcz6q") pod "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" (UID: "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8"). InnerVolumeSpecName "kube-api-access-xcz6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.451858 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-config-data" (OuterVolumeSpecName: "config-data") pod "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" (UID: "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.453474 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" (UID: "4aa2367d-dad7-461e-9e93-1ebc2bfee6e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.524460 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcz6q\" (UniqueName: \"kubernetes.io/projected/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-kube-api-access-xcz6q\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.524501 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.524515 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.524527 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.919007 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2pxck" event={"ID":"4aa2367d-dad7-461e-9e93-1ebc2bfee6e8","Type":"ContainerDied","Data":"6305823f18f4277ffc6bc86133f2d62ee65779ea4a52d25d602c9e38aaa1b86c"} Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.919337 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6305823f18f4277ffc6bc86133f2d62ee65779ea4a52d25d602c9e38aaa1b86c" Dec 05 11:32:05 crc kubenswrapper[4809]: I1205 11:32:05.919114 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2pxck" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.066996 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:32:06 crc kubenswrapper[4809]: E1205 11:32:06.067766 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" containerName="nova-cell0-conductor-db-sync" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.067796 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" containerName="nova-cell0-conductor-db-sync" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.068281 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" containerName="nova-cell0-conductor-db-sync" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.070339 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.072956 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fc49v" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.079432 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.081965 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.138046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.138095 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8n6k\" (UniqueName: \"kubernetes.io/projected/23b3a18d-30f0-496f-9990-dd453ea8dbb7-kube-api-access-p8n6k\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.138121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.239954 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.240013 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8n6k\" (UniqueName: \"kubernetes.io/projected/23b3a18d-30f0-496f-9990-dd453ea8dbb7-kube-api-access-p8n6k\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.240038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.249749 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.251213 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.260987 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8n6k\" (UniqueName: \"kubernetes.io/projected/23b3a18d-30f0-496f-9990-dd453ea8dbb7-kube-api-access-p8n6k\") pod \"nova-cell0-conductor-0\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.391793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.838792 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:32:06 crc kubenswrapper[4809]: I1205 11:32:06.934409 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"23b3a18d-30f0-496f-9990-dd453ea8dbb7","Type":"ContainerStarted","Data":"c0a31ba61b268e35d82eeb2c4628dc61e5cfaa9fe1ee7bc86d26ed72479bcf4d"} Dec 05 11:32:08 crc kubenswrapper[4809]: I1205 11:32:08.979512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"23b3a18d-30f0-496f-9990-dd453ea8dbb7","Type":"ContainerStarted","Data":"57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a"} Dec 05 11:32:08 crc kubenswrapper[4809]: I1205 11:32:08.979908 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:09 crc kubenswrapper[4809]: I1205 11:32:09.007291 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.007275638 podStartE2EDuration="3.007275638s" podCreationTimestamp="2025-12-05 11:32:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:08.999292449 +0000 UTC m=+1424.390269007" watchObservedRunningTime="2025-12-05 11:32:09.007275638 +0000 UTC m=+1424.398252186" Dec 05 11:32:16 crc kubenswrapper[4809]: I1205 11:32:16.420751 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 11:32:16 crc kubenswrapper[4809]: I1205 11:32:16.909733 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fqflm"] Dec 05 11:32:16 crc kubenswrapper[4809]: I1205 11:32:16.915489 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:16 crc kubenswrapper[4809]: I1205 11:32:16.915565 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fqflm"] Dec 05 11:32:16 crc kubenswrapper[4809]: I1205 11:32:16.950999 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 05 11:32:16 crc kubenswrapper[4809]: I1205 11:32:16.951343 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.062873 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.064296 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.067268 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.072328 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.072458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-scripts\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.072481 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwx8w\" (UniqueName: \"kubernetes.io/projected/0b7d8ff5-5828-4f49-9400-bdb47aa56419-kube-api-access-cwx8w\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.072501 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-config-data\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.093201 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.139280 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.140736 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.142695 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.156222 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.175497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.175975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.176164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9x9w\" (UniqueName: \"kubernetes.io/projected/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-kube-api-access-r9x9w\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.176298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.176379 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-scripts\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.176476 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwx8w\" (UniqueName: \"kubernetes.io/projected/0b7d8ff5-5828-4f49-9400-bdb47aa56419-kube-api-access-cwx8w\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.176556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-config-data\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.184755 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.187413 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-config-data\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.206561 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwx8w\" (UniqueName: \"kubernetes.io/projected/0b7d8ff5-5828-4f49-9400-bdb47aa56419-kube-api-access-cwx8w\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.207164 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-scripts\") pod \"nova-cell0-cell-mapping-fqflm\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.259979 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.280160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8s7r\" (UniqueName: \"kubernetes.io/projected/89b7fd8e-9acc-4069-a652-b033f142cc6c-kube-api-access-q8s7r\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.280221 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.280249 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.280304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b7fd8e-9acc-4069-a652-b033f142cc6c-logs\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.280356 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9x9w\" (UniqueName: \"kubernetes.io/projected/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-kube-api-access-r9x9w\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.280544 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.280582 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-config-data\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.295231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.296200 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.297994 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.300985 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.305220 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.318096 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9x9w\" (UniqueName: \"kubernetes.io/projected/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-kube-api-access-r9x9w\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.324101 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.358706 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.359849 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.368447 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386387 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg9kx\" (UniqueName: \"kubernetes.io/projected/b71880e0-2df6-4c3b-8567-eeb60a9e7965-kube-api-access-vg9kx\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386408 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b7fd8e-9acc-4069-a652-b033f142cc6c-logs\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386456 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71880e0-2df6-4c3b-8567-eeb60a9e7965-logs\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386479 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386546 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-config-data\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.386578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8s7r\" (UniqueName: \"kubernetes.io/projected/89b7fd8e-9acc-4069-a652-b033f142cc6c-kube-api-access-q8s7r\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.387940 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b7fd8e-9acc-4069-a652-b033f142cc6c-logs\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.388517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.400696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.407282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-config-data\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.409767 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.445772 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8s7r\" (UniqueName: \"kubernetes.io/projected/89b7fd8e-9acc-4069-a652-b033f142cc6c-kube-api-access-q8s7r\") pod \"nova-api-0\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.447855 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-2pv8s"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.449356 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.455653 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-2pv8s"] Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.474569 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.487863 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.487953 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.487983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-config-data\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.488000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdq64\" (UniqueName: \"kubernetes.io/projected/feccc379-3fb9-4271-a582-3412618bfd12-kube-api-access-sdq64\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.488032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg9kx\" (UniqueName: \"kubernetes.io/projected/b71880e0-2df6-4c3b-8567-eeb60a9e7965-kube-api-access-vg9kx\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.488291 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71880e0-2df6-4c3b-8567-eeb60a9e7965-logs\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.488339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.491337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71880e0-2df6-4c3b-8567-eeb60a9e7965-logs\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.496541 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.505340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.508395 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg9kx\" (UniqueName: \"kubernetes.io/projected/b71880e0-2df6-4c3b-8567-eeb60a9e7965-kube-api-access-vg9kx\") pod \"nova-metadata-0\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589697 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589752 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvlrg\" (UniqueName: \"kubernetes.io/projected/af9a1e65-a606-4221-96e2-72e11675fd31-kube-api-access-cvlrg\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589800 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-config-data\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589874 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-config\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589914 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdq64\" (UniqueName: \"kubernetes.io/projected/feccc379-3fb9-4271-a582-3412618bfd12-kube-api-access-sdq64\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.589996 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.604177 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-config-data\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.605652 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.629385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdq64\" (UniqueName: \"kubernetes.io/projected/feccc379-3fb9-4271-a582-3412618bfd12-kube-api-access-sdq64\") pod \"nova-scheduler-0\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.693261 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.693856 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvlrg\" (UniqueName: \"kubernetes.io/projected/af9a1e65-a606-4221-96e2-72e11675fd31-kube-api-access-cvlrg\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.693908 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.693968 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-config\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.694084 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.694119 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.694373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.695643 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.695743 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-config\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.696260 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.696289 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.714142 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvlrg\" (UniqueName: \"kubernetes.io/projected/af9a1e65-a606-4221-96e2-72e11675fd31-kube-api-access-cvlrg\") pod \"dnsmasq-dns-845d6d6f59-2pv8s\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.771150 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.787396 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.798970 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:17 crc kubenswrapper[4809]: I1205 11:32:17.983450 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fqflm"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.067742 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.081218 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.107871 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fqflm" event={"ID":"0b7d8ff5-5828-4f49-9400-bdb47aa56419","Type":"ContainerStarted","Data":"487b13579819ee13743adf76309739ff77962f32c815a9784efd1cf01a7456b6"} Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.128241 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l9n4x"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.129583 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.133615 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.133877 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.142914 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l9n4x"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.231720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.231775 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg7s2\" (UniqueName: \"kubernetes.io/projected/5ad0cc4c-461f-4de2-aaf0-659c655edff3-kube-api-access-pg7s2\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.231809 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-scripts\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.231837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-config-data\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.333860 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.334186 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg7s2\" (UniqueName: \"kubernetes.io/projected/5ad0cc4c-461f-4de2-aaf0-659c655edff3-kube-api-access-pg7s2\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.334214 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-scripts\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.334238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-config-data\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.340044 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-config-data\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.340366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.344019 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-scripts\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.360316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg7s2\" (UniqueName: \"kubernetes.io/projected/5ad0cc4c-461f-4de2-aaf0-659c655edff3-kube-api-access-pg7s2\") pod \"nova-cell1-conductor-db-sync-l9n4x\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.425747 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-2pv8s"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.430306 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.471935 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.574160 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:18 crc kubenswrapper[4809]: I1205 11:32:18.963504 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l9n4x"] Dec 05 11:32:18 crc kubenswrapper[4809]: W1205 11:32:18.980258 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ad0cc4c_461f_4de2_aaf0_659c655edff3.slice/crio-4e2aad20458ad621cd4ae6a1c8991ca350824711aa30eca65469689d1430ec34 WatchSource:0}: Error finding container 4e2aad20458ad621cd4ae6a1c8991ca350824711aa30eca65469689d1430ec34: Status 404 returned error can't find the container with id 4e2aad20458ad621cd4ae6a1c8991ca350824711aa30eca65469689d1430ec34 Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.128822 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"feccc379-3fb9-4271-a582-3412618bfd12","Type":"ContainerStarted","Data":"50f848b9e33217ba3011da326b729463e22c8f854b8e7c86e533d6bbbe433406"} Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.131010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" event={"ID":"5ad0cc4c-461f-4de2-aaf0-659c655edff3","Type":"ContainerStarted","Data":"4e2aad20458ad621cd4ae6a1c8991ca350824711aa30eca65469689d1430ec34"} Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.139331 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b71880e0-2df6-4c3b-8567-eeb60a9e7965","Type":"ContainerStarted","Data":"90dc2422484fc3c7738287f941236e645ec01b5bcd8776e2f46dd231cf2c25a6"} Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.150971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"89b7fd8e-9acc-4069-a652-b033f142cc6c","Type":"ContainerStarted","Data":"446aee61012549036f10012e22f2fd02e291178ea5aaa335150a423d74e8fab4"} Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.161980 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fqflm" event={"ID":"0b7d8ff5-5828-4f49-9400-bdb47aa56419","Type":"ContainerStarted","Data":"9af9fd7f8d169447279e6fa340bbf6d3969e9c9fff1059f5d537b02471d23b2d"} Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.167396 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf","Type":"ContainerStarted","Data":"afb10f41a5a60585b85d154b55a42fc59b750d7d1adb3cd93fafbdfb077dea25"} Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.186264 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fqflm" podStartSLOduration=3.186245333 podStartE2EDuration="3.186245333s" podCreationTimestamp="2025-12-05 11:32:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:19.179302483 +0000 UTC m=+1434.570279041" watchObservedRunningTime="2025-12-05 11:32:19.186245333 +0000 UTC m=+1434.577221891" Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.200727 4809 generic.go:334] "Generic (PLEG): container finished" podID="af9a1e65-a606-4221-96e2-72e11675fd31" containerID="b6e1b17faedc5097d9b2d6c223245dc1c85527026a411fd27c7c506de3918ad2" exitCode=0 Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.200785 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" event={"ID":"af9a1e65-a606-4221-96e2-72e11675fd31","Type":"ContainerDied","Data":"b6e1b17faedc5097d9b2d6c223245dc1c85527026a411fd27c7c506de3918ad2"} Dec 05 11:32:19 crc kubenswrapper[4809]: I1205 11:32:19.200809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" event={"ID":"af9a1e65-a606-4221-96e2-72e11675fd31","Type":"ContainerStarted","Data":"2b22c1f28c422cc5d8682b3df064105ef903bde29290ffb106a21fef064706d5"} Dec 05 11:32:20 crc kubenswrapper[4809]: I1205 11:32:20.213734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" event={"ID":"af9a1e65-a606-4221-96e2-72e11675fd31","Type":"ContainerStarted","Data":"7263bb11bc485b26900d63cf7d94328801219440e7d4e00a3937691228bd52c7"} Dec 05 11:32:20 crc kubenswrapper[4809]: I1205 11:32:20.214132 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:20 crc kubenswrapper[4809]: I1205 11:32:20.219845 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" event={"ID":"5ad0cc4c-461f-4de2-aaf0-659c655edff3","Type":"ContainerStarted","Data":"d66ed99c8c65211097d15a89bde5605afdc64db5d12e59bb72da9739aeaa7f01"} Dec 05 11:32:20 crc kubenswrapper[4809]: I1205 11:32:20.240569 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" podStartSLOduration=3.240547753 podStartE2EDuration="3.240547753s" podCreationTimestamp="2025-12-05 11:32:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:20.239306891 +0000 UTC m=+1435.630283469" watchObservedRunningTime="2025-12-05 11:32:20.240547753 +0000 UTC m=+1435.631524311" Dec 05 11:32:20 crc kubenswrapper[4809]: I1205 11:32:20.255924 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" podStartSLOduration=2.255905449 podStartE2EDuration="2.255905449s" podCreationTimestamp="2025-12-05 11:32:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:20.254492633 +0000 UTC m=+1435.645469191" watchObservedRunningTime="2025-12-05 11:32:20.255905449 +0000 UTC m=+1435.646882027" Dec 05 11:32:21 crc kubenswrapper[4809]: I1205 11:32:21.254095 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:21 crc kubenswrapper[4809]: I1205 11:32:21.281097 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.212870 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.272182 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b71880e0-2df6-4c3b-8567-eeb60a9e7965","Type":"ContainerStarted","Data":"70d9e01fef5bc94f733f102d230195a4412e21f1cd42d3ca7f642696d067fdc9"} Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.285886 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"89b7fd8e-9acc-4069-a652-b033f142cc6c","Type":"ContainerStarted","Data":"f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6"} Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.292757 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf","Type":"ContainerStarted","Data":"894c6277b8486291c59a23e96d6969fc19f092834afc5ce802d6b8dc9d69bcc6"} Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.292896 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://894c6277b8486291c59a23e96d6969fc19f092834afc5ce802d6b8dc9d69bcc6" gracePeriod=30 Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.296177 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"feccc379-3fb9-4271-a582-3412618bfd12","Type":"ContainerStarted","Data":"0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a"} Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.320658 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.716490303 podStartE2EDuration="5.32062041s" podCreationTimestamp="2025-12-05 11:32:17 +0000 UTC" firstStartedPulling="2025-12-05 11:32:18.106922786 +0000 UTC m=+1433.497899354" lastFinishedPulling="2025-12-05 11:32:21.711052903 +0000 UTC m=+1437.102029461" observedRunningTime="2025-12-05 11:32:22.309711398 +0000 UTC m=+1437.700687956" watchObservedRunningTime="2025-12-05 11:32:22.32062041 +0000 UTC m=+1437.711596978" Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.389568 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:22 crc kubenswrapper[4809]: I1205 11:32:22.788475 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 11:32:23 crc kubenswrapper[4809]: I1205 11:32:23.315997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b71880e0-2df6-4c3b-8567-eeb60a9e7965","Type":"ContainerStarted","Data":"758b687f72a2c6fac5cc1b865c6bbb9a86a4664ec4f448e4631547c3a0aba1a9"} Dec 05 11:32:23 crc kubenswrapper[4809]: I1205 11:32:23.316215 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-log" containerID="cri-o://70d9e01fef5bc94f733f102d230195a4412e21f1cd42d3ca7f642696d067fdc9" gracePeriod=30 Dec 05 11:32:23 crc kubenswrapper[4809]: I1205 11:32:23.317160 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-metadata" containerID="cri-o://758b687f72a2c6fac5cc1b865c6bbb9a86a4664ec4f448e4631547c3a0aba1a9" gracePeriod=30 Dec 05 11:32:23 crc kubenswrapper[4809]: I1205 11:32:23.324470 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"89b7fd8e-9acc-4069-a652-b033f142cc6c","Type":"ContainerStarted","Data":"ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59"} Dec 05 11:32:23 crc kubenswrapper[4809]: I1205 11:32:23.346590 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.09081576 podStartE2EDuration="6.346572659s" podCreationTimestamp="2025-12-05 11:32:17 +0000 UTC" firstStartedPulling="2025-12-05 11:32:18.459466801 +0000 UTC m=+1433.850443359" lastFinishedPulling="2025-12-05 11:32:21.7152237 +0000 UTC m=+1437.106200258" observedRunningTime="2025-12-05 11:32:23.345376248 +0000 UTC m=+1438.736352816" watchObservedRunningTime="2025-12-05 11:32:23.346572659 +0000 UTC m=+1438.737549217" Dec 05 11:32:23 crc kubenswrapper[4809]: I1205 11:32:23.352312 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.241233191 podStartE2EDuration="6.352300677s" podCreationTimestamp="2025-12-05 11:32:17 +0000 UTC" firstStartedPulling="2025-12-05 11:32:18.5981617 +0000 UTC m=+1433.989138258" lastFinishedPulling="2025-12-05 11:32:21.709229186 +0000 UTC m=+1437.100205744" observedRunningTime="2025-12-05 11:32:22.33690294 +0000 UTC m=+1437.727879498" watchObservedRunningTime="2025-12-05 11:32:23.352300677 +0000 UTC m=+1438.743277235" Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.334104 4809 generic.go:334] "Generic (PLEG): container finished" podID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerID="758b687f72a2c6fac5cc1b865c6bbb9a86a4664ec4f448e4631547c3a0aba1a9" exitCode=0 Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.334429 4809 generic.go:334] "Generic (PLEG): container finished" podID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerID="70d9e01fef5bc94f733f102d230195a4412e21f1cd42d3ca7f642696d067fdc9" exitCode=143 Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.334786 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b71880e0-2df6-4c3b-8567-eeb60a9e7965","Type":"ContainerDied","Data":"758b687f72a2c6fac5cc1b865c6bbb9a86a4664ec4f448e4631547c3a0aba1a9"} Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.334856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b71880e0-2df6-4c3b-8567-eeb60a9e7965","Type":"ContainerDied","Data":"70d9e01fef5bc94f733f102d230195a4412e21f1cd42d3ca7f642696d067fdc9"} Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.470126 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.494480 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.899019762 podStartE2EDuration="7.494457335s" podCreationTimestamp="2025-12-05 11:32:17 +0000 UTC" firstStartedPulling="2025-12-05 11:32:18.115595329 +0000 UTC m=+1433.506571897" lastFinishedPulling="2025-12-05 11:32:21.711032912 +0000 UTC m=+1437.102009470" observedRunningTime="2025-12-05 11:32:23.375168937 +0000 UTC m=+1438.766145515" watchObservedRunningTime="2025-12-05 11:32:24.494457335 +0000 UTC m=+1439.885433913" Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.593291 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71880e0-2df6-4c3b-8567-eeb60a9e7965-logs\") pod \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.593391 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data\") pod \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.593680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-combined-ca-bundle\") pod \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.593762 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg9kx\" (UniqueName: \"kubernetes.io/projected/b71880e0-2df6-4c3b-8567-eeb60a9e7965-kube-api-access-vg9kx\") pod \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.593975 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b71880e0-2df6-4c3b-8567-eeb60a9e7965-logs" (OuterVolumeSpecName: "logs") pod "b71880e0-2df6-4c3b-8567-eeb60a9e7965" (UID: "b71880e0-2df6-4c3b-8567-eeb60a9e7965"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.594668 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b71880e0-2df6-4c3b-8567-eeb60a9e7965-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.600806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b71880e0-2df6-4c3b-8567-eeb60a9e7965-kube-api-access-vg9kx" (OuterVolumeSpecName: "kube-api-access-vg9kx") pod "b71880e0-2df6-4c3b-8567-eeb60a9e7965" (UID: "b71880e0-2df6-4c3b-8567-eeb60a9e7965"). InnerVolumeSpecName "kube-api-access-vg9kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:24 crc kubenswrapper[4809]: E1205 11:32:24.625202 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data podName:b71880e0-2df6-4c3b-8567-eeb60a9e7965 nodeName:}" failed. No retries permitted until 2025-12-05 11:32:25.125156657 +0000 UTC m=+1440.516133215 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data") pod "b71880e0-2df6-4c3b-8567-eeb60a9e7965" (UID: "b71880e0-2df6-4c3b-8567-eeb60a9e7965") : error deleting /var/lib/kubelet/pods/b71880e0-2df6-4c3b-8567-eeb60a9e7965/volume-subpaths: remove /var/lib/kubelet/pods/b71880e0-2df6-4c3b-8567-eeb60a9e7965/volume-subpaths: no such file or directory Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.630499 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b71880e0-2df6-4c3b-8567-eeb60a9e7965" (UID: "b71880e0-2df6-4c3b-8567-eeb60a9e7965"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.697337 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:24 crc kubenswrapper[4809]: I1205 11:32:24.697397 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg9kx\" (UniqueName: \"kubernetes.io/projected/b71880e0-2df6-4c3b-8567-eeb60a9e7965-kube-api-access-vg9kx\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.207416 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data\") pod \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\" (UID: \"b71880e0-2df6-4c3b-8567-eeb60a9e7965\") " Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.217056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data" (OuterVolumeSpecName: "config-data") pod "b71880e0-2df6-4c3b-8567-eeb60a9e7965" (UID: "b71880e0-2df6-4c3b-8567-eeb60a9e7965"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.310450 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b71880e0-2df6-4c3b-8567-eeb60a9e7965-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.351412 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b71880e0-2df6-4c3b-8567-eeb60a9e7965","Type":"ContainerDied","Data":"90dc2422484fc3c7738287f941236e645ec01b5bcd8776e2f46dd231cf2c25a6"} Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.351684 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.351853 4809 scope.go:117] "RemoveContainer" containerID="758b687f72a2c6fac5cc1b865c6bbb9a86a4664ec4f448e4631547c3a0aba1a9" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.390094 4809 scope.go:117] "RemoveContainer" containerID="70d9e01fef5bc94f733f102d230195a4412e21f1cd42d3ca7f642696d067fdc9" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.402718 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.411408 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.439319 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:25 crc kubenswrapper[4809]: E1205 11:32:25.439911 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-metadata" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.439937 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-metadata" Dec 05 11:32:25 crc kubenswrapper[4809]: E1205 11:32:25.439984 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-log" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.439995 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-log" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.440319 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-log" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.440368 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" containerName="nova-metadata-metadata" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.441772 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.445154 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.448538 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.451991 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.618770 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef450f73-b283-4a71-82ca-f93bbdfc31dc-logs\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.618848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.618940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.619095 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24gmd\" (UniqueName: \"kubernetes.io/projected/ef450f73-b283-4a71-82ca-f93bbdfc31dc-kube-api-access-24gmd\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.619192 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-config-data\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.720749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.720838 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24gmd\" (UniqueName: \"kubernetes.io/projected/ef450f73-b283-4a71-82ca-f93bbdfc31dc-kube-api-access-24gmd\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.720906 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-config-data\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.720998 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef450f73-b283-4a71-82ca-f93bbdfc31dc-logs\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.721022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.722143 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef450f73-b283-4a71-82ca-f93bbdfc31dc-logs\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.752205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.752574 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.752738 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-config-data\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.756467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24gmd\" (UniqueName: \"kubernetes.io/projected/ef450f73-b283-4a71-82ca-f93bbdfc31dc-kube-api-access-24gmd\") pod \"nova-metadata-0\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " pod="openstack/nova-metadata-0" Dec 05 11:32:25 crc kubenswrapper[4809]: I1205 11:32:25.772423 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:26 crc kubenswrapper[4809]: I1205 11:32:26.217936 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:26 crc kubenswrapper[4809]: W1205 11:32:26.224407 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef450f73_b283_4a71_82ca_f93bbdfc31dc.slice/crio-99531f22858cedd596174b351bf538b67f8cf4247ad376b9513538e7e4fced79 WatchSource:0}: Error finding container 99531f22858cedd596174b351bf538b67f8cf4247ad376b9513538e7e4fced79: Status 404 returned error can't find the container with id 99531f22858cedd596174b351bf538b67f8cf4247ad376b9513538e7e4fced79 Dec 05 11:32:26 crc kubenswrapper[4809]: I1205 11:32:26.369006 4809 generic.go:334] "Generic (PLEG): container finished" podID="0b7d8ff5-5828-4f49-9400-bdb47aa56419" containerID="9af9fd7f8d169447279e6fa340bbf6d3969e9c9fff1059f5d537b02471d23b2d" exitCode=0 Dec 05 11:32:26 crc kubenswrapper[4809]: I1205 11:32:26.369077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fqflm" event={"ID":"0b7d8ff5-5828-4f49-9400-bdb47aa56419","Type":"ContainerDied","Data":"9af9fd7f8d169447279e6fa340bbf6d3969e9c9fff1059f5d537b02471d23b2d"} Dec 05 11:32:26 crc kubenswrapper[4809]: I1205 11:32:26.370742 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef450f73-b283-4a71-82ca-f93bbdfc31dc","Type":"ContainerStarted","Data":"99531f22858cedd596174b351bf538b67f8cf4247ad376b9513538e7e4fced79"} Dec 05 11:32:26 crc kubenswrapper[4809]: I1205 11:32:26.897298 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b71880e0-2df6-4c3b-8567-eeb60a9e7965" path="/var/lib/kubelet/pods/b71880e0-2df6-4c3b-8567-eeb60a9e7965/volumes" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.383572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef450f73-b283-4a71-82ca-f93bbdfc31dc","Type":"ContainerStarted","Data":"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832"} Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.384062 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef450f73-b283-4a71-82ca-f93bbdfc31dc","Type":"ContainerStarted","Data":"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c"} Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.411259 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.411234258 podStartE2EDuration="2.411234258s" podCreationTimestamp="2025-12-05 11:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:27.404785471 +0000 UTC m=+1442.795762079" watchObservedRunningTime="2025-12-05 11:32:27.411234258 +0000 UTC m=+1442.802210816" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.475407 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.475927 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.783088 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.788336 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.800753 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.849522 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.862810 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-8h6tp"] Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.862868 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-config-data\") pod \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.862975 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-scripts\") pod \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.863060 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" podUID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerName="dnsmasq-dns" containerID="cri-o://41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4" gracePeriod=10 Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.863083 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwx8w\" (UniqueName: \"kubernetes.io/projected/0b7d8ff5-5828-4f49-9400-bdb47aa56419-kube-api-access-cwx8w\") pod \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.863134 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-combined-ca-bundle\") pod \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\" (UID: \"0b7d8ff5-5828-4f49-9400-bdb47aa56419\") " Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.880308 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7d8ff5-5828-4f49-9400-bdb47aa56419-kube-api-access-cwx8w" (OuterVolumeSpecName: "kube-api-access-cwx8w") pod "0b7d8ff5-5828-4f49-9400-bdb47aa56419" (UID: "0b7d8ff5-5828-4f49-9400-bdb47aa56419"). InnerVolumeSpecName "kube-api-access-cwx8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.886193 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-scripts" (OuterVolumeSpecName: "scripts") pod "0b7d8ff5-5828-4f49-9400-bdb47aa56419" (UID: "0b7d8ff5-5828-4f49-9400-bdb47aa56419"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.932058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b7d8ff5-5828-4f49-9400-bdb47aa56419" (UID: "0b7d8ff5-5828-4f49-9400-bdb47aa56419"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.957262 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-config-data" (OuterVolumeSpecName: "config-data") pod "0b7d8ff5-5828-4f49-9400-bdb47aa56419" (UID: "0b7d8ff5-5828-4f49-9400-bdb47aa56419"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.966600 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.966653 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.966666 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwx8w\" (UniqueName: \"kubernetes.io/projected/0b7d8ff5-5828-4f49-9400-bdb47aa56419-kube-api-access-cwx8w\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:27 crc kubenswrapper[4809]: I1205 11:32:27.966682 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7d8ff5-5828-4f49-9400-bdb47aa56419-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.320909 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.395355 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ad0cc4c-461f-4de2-aaf0-659c655edff3" containerID="d66ed99c8c65211097d15a89bde5605afdc64db5d12e59bb72da9739aeaa7f01" exitCode=0 Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.395422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" event={"ID":"5ad0cc4c-461f-4de2-aaf0-659c655edff3","Type":"ContainerDied","Data":"d66ed99c8c65211097d15a89bde5605afdc64db5d12e59bb72da9739aeaa7f01"} Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.415431 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fqflm" event={"ID":"0b7d8ff5-5828-4f49-9400-bdb47aa56419","Type":"ContainerDied","Data":"487b13579819ee13743adf76309739ff77962f32c815a9784efd1cf01a7456b6"} Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.415467 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="487b13579819ee13743adf76309739ff77962f32c815a9784efd1cf01a7456b6" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.415540 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fqflm" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.417312 4809 generic.go:334] "Generic (PLEG): container finished" podID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerID="41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4" exitCode=0 Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.418061 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.418430 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" event={"ID":"d52c2032-df32-4ff3-ab67-3c623e6fe80c","Type":"ContainerDied","Data":"41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4"} Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.418451 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-8h6tp" event={"ID":"d52c2032-df32-4ff3-ab67-3c623e6fe80c","Type":"ContainerDied","Data":"141e7c8d0fe6b843b8eef6a5be1fdfccb90c019b85d296aa18caf1f74fa6f2bb"} Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.418468 4809 scope.go:117] "RemoveContainer" containerID="41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.453528 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.456889 4809 scope.go:117] "RemoveContainer" containerID="88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.485939 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-svc\") pod \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.486048 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-sb\") pod \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.486075 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-nb\") pod \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.486104 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq6fh\" (UniqueName: \"kubernetes.io/projected/d52c2032-df32-4ff3-ab67-3c623e6fe80c-kube-api-access-jq6fh\") pod \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.486196 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-swift-storage-0\") pod \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.486271 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-config\") pod \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\" (UID: \"d52c2032-df32-4ff3-ab67-3c623e6fe80c\") " Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.497329 4809 scope.go:117] "RemoveContainer" containerID="41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.500794 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d52c2032-df32-4ff3-ab67-3c623e6fe80c-kube-api-access-jq6fh" (OuterVolumeSpecName: "kube-api-access-jq6fh") pod "d52c2032-df32-4ff3-ab67-3c623e6fe80c" (UID: "d52c2032-df32-4ff3-ab67-3c623e6fe80c"). InnerVolumeSpecName "kube-api-access-jq6fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:28 crc kubenswrapper[4809]: E1205 11:32:28.502884 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4\": container with ID starting with 41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4 not found: ID does not exist" containerID="41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.502946 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4"} err="failed to get container status \"41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4\": rpc error: code = NotFound desc = could not find container \"41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4\": container with ID starting with 41d88e82b84ead3be3e219609d14bd3fc6744dfe3d6b6b7ef17a113c56953fc4 not found: ID does not exist" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.502976 4809 scope.go:117] "RemoveContainer" containerID="88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32" Dec 05 11:32:28 crc kubenswrapper[4809]: E1205 11:32:28.503395 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32\": container with ID starting with 88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32 not found: ID does not exist" containerID="88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.503418 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32"} err="failed to get container status \"88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32\": rpc error: code = NotFound desc = could not find container \"88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32\": container with ID starting with 88026eadc107cd4b7a7c8fb801e4bd050777533c49e21f5b38ce26a30a3aca32 not found: ID does not exist" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.552150 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.564240 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.565376 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-config" (OuterVolumeSpecName: "config") pod "d52c2032-df32-4ff3-ab67-3c623e6fe80c" (UID: "d52c2032-df32-4ff3-ab67-3c623e6fe80c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.567242 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.567578 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.586480 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d52c2032-df32-4ff3-ab67-3c623e6fe80c" (UID: "d52c2032-df32-4ff3-ab67-3c623e6fe80c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.588126 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.588155 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq6fh\" (UniqueName: \"kubernetes.io/projected/d52c2032-df32-4ff3-ab67-3c623e6fe80c-kube-api-access-jq6fh\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.588167 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.596142 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d52c2032-df32-4ff3-ab67-3c623e6fe80c" (UID: "d52c2032-df32-4ff3-ab67-3c623e6fe80c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.596652 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d52c2032-df32-4ff3-ab67-3c623e6fe80c" (UID: "d52c2032-df32-4ff3-ab67-3c623e6fe80c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.598546 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d52c2032-df32-4ff3-ab67-3c623e6fe80c" (UID: "d52c2032-df32-4ff3-ab67-3c623e6fe80c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.690016 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.690050 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.690059 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d52c2032-df32-4ff3-ab67-3c623e6fe80c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.748395 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-8h6tp"] Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.778448 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-8h6tp"] Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.901983 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" path="/var/lib/kubelet/pods/d52c2032-df32-4ff3-ab67-3c623e6fe80c/volumes" Dec 05 11:32:28 crc kubenswrapper[4809]: I1205 11:32:28.969117 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:29 crc kubenswrapper[4809]: I1205 11:32:29.427751 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-log" containerID="cri-o://1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c" gracePeriod=30 Dec 05 11:32:29 crc kubenswrapper[4809]: I1205 11:32:29.427868 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-metadata" containerID="cri-o://8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832" gracePeriod=30 Dec 05 11:32:29 crc kubenswrapper[4809]: I1205 11:32:29.428110 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-api" containerID="cri-o://ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59" gracePeriod=30 Dec 05 11:32:29 crc kubenswrapper[4809]: I1205 11:32:29.428053 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-log" containerID="cri-o://f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6" gracePeriod=30 Dec 05 11:32:29 crc kubenswrapper[4809]: I1205 11:32:29.851037 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.020323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg7s2\" (UniqueName: \"kubernetes.io/projected/5ad0cc4c-461f-4de2-aaf0-659c655edff3-kube-api-access-pg7s2\") pod \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.020396 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-config-data\") pod \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.020672 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-combined-ca-bundle\") pod \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.020725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-scripts\") pod \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\" (UID: \"5ad0cc4c-461f-4de2-aaf0-659c655edff3\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.032478 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad0cc4c-461f-4de2-aaf0-659c655edff3-kube-api-access-pg7s2" (OuterVolumeSpecName: "kube-api-access-pg7s2") pod "5ad0cc4c-461f-4de2-aaf0-659c655edff3" (UID: "5ad0cc4c-461f-4de2-aaf0-659c655edff3"). InnerVolumeSpecName "kube-api-access-pg7s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.050926 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ad0cc4c-461f-4de2-aaf0-659c655edff3" (UID: "5ad0cc4c-461f-4de2-aaf0-659c655edff3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.054785 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-scripts" (OuterVolumeSpecName: "scripts") pod "5ad0cc4c-461f-4de2-aaf0-659c655edff3" (UID: "5ad0cc4c-461f-4de2-aaf0-659c655edff3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.097924 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.131333 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.131602 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.131831 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg7s2\" (UniqueName: \"kubernetes.io/projected/5ad0cc4c-461f-4de2-aaf0-659c655edff3-kube-api-access-pg7s2\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.184838 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-config-data" (OuterVolumeSpecName: "config-data") pod "5ad0cc4c-461f-4de2-aaf0-659c655edff3" (UID: "5ad0cc4c-461f-4de2-aaf0-659c655edff3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.232762 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef450f73-b283-4a71-82ca-f93bbdfc31dc-logs\") pod \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.232870 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-combined-ca-bundle\") pod \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.232953 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24gmd\" (UniqueName: \"kubernetes.io/projected/ef450f73-b283-4a71-82ca-f93bbdfc31dc-kube-api-access-24gmd\") pod \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.233138 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-config-data\") pod \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.233174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-nova-metadata-tls-certs\") pod \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\" (UID: \"ef450f73-b283-4a71-82ca-f93bbdfc31dc\") " Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.233193 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef450f73-b283-4a71-82ca-f93bbdfc31dc-logs" (OuterVolumeSpecName: "logs") pod "ef450f73-b283-4a71-82ca-f93bbdfc31dc" (UID: "ef450f73-b283-4a71-82ca-f93bbdfc31dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.233717 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef450f73-b283-4a71-82ca-f93bbdfc31dc-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.233738 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad0cc4c-461f-4de2-aaf0-659c655edff3-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.257367 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef450f73-b283-4a71-82ca-f93bbdfc31dc-kube-api-access-24gmd" (OuterVolumeSpecName: "kube-api-access-24gmd") pod "ef450f73-b283-4a71-82ca-f93bbdfc31dc" (UID: "ef450f73-b283-4a71-82ca-f93bbdfc31dc"). InnerVolumeSpecName "kube-api-access-24gmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.260437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef450f73-b283-4a71-82ca-f93bbdfc31dc" (UID: "ef450f73-b283-4a71-82ca-f93bbdfc31dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.280947 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-config-data" (OuterVolumeSpecName: "config-data") pod "ef450f73-b283-4a71-82ca-f93bbdfc31dc" (UID: "ef450f73-b283-4a71-82ca-f93bbdfc31dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.303201 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ef450f73-b283-4a71-82ca-f93bbdfc31dc" (UID: "ef450f73-b283-4a71-82ca-f93bbdfc31dc"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.335748 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.335968 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24gmd\" (UniqueName: \"kubernetes.io/projected/ef450f73-b283-4a71-82ca-f93bbdfc31dc-kube-api-access-24gmd\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.336028 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.336083 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef450f73-b283-4a71-82ca-f93bbdfc31dc-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.440799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" event={"ID":"5ad0cc4c-461f-4de2-aaf0-659c655edff3","Type":"ContainerDied","Data":"4e2aad20458ad621cd4ae6a1c8991ca350824711aa30eca65469689d1430ec34"} Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.441484 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e2aad20458ad621cd4ae6a1c8991ca350824711aa30eca65469689d1430ec34" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.441608 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l9n4x" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.450173 4809 generic.go:334] "Generic (PLEG): container finished" podID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerID="f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6" exitCode=143 Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.450243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"89b7fd8e-9acc-4069-a652-b033f142cc6c","Type":"ContainerDied","Data":"f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6"} Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452048 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerID="8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832" exitCode=0 Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452073 4809 generic.go:334] "Generic (PLEG): container finished" podID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerID="1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c" exitCode=143 Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452214 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="feccc379-3fb9-4271-a582-3412618bfd12" containerName="nova-scheduler-scheduler" containerID="cri-o://0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" gracePeriod=30 Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef450f73-b283-4a71-82ca-f93bbdfc31dc","Type":"ContainerDied","Data":"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832"} Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452287 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef450f73-b283-4a71-82ca-f93bbdfc31dc","Type":"ContainerDied","Data":"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c"} Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ef450f73-b283-4a71-82ca-f93bbdfc31dc","Type":"ContainerDied","Data":"99531f22858cedd596174b351bf538b67f8cf4247ad376b9513538e7e4fced79"} Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.452342 4809 scope.go:117] "RemoveContainer" containerID="8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.476810 4809 scope.go:117] "RemoveContainer" containerID="1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.497934 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.512070 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.515153 4809 scope.go:117] "RemoveContainer" containerID="8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832" Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.515620 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832\": container with ID starting with 8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832 not found: ID does not exist" containerID="8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.515731 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832"} err="failed to get container status \"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832\": rpc error: code = NotFound desc = could not find container \"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832\": container with ID starting with 8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832 not found: ID does not exist" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.515810 4809 scope.go:117] "RemoveContainer" containerID="1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c" Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.516356 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c\": container with ID starting with 1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c not found: ID does not exist" containerID="1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.516429 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c"} err="failed to get container status \"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c\": rpc error: code = NotFound desc = could not find container \"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c\": container with ID starting with 1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c not found: ID does not exist" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.516499 4809 scope.go:117] "RemoveContainer" containerID="8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.517030 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832"} err="failed to get container status \"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832\": rpc error: code = NotFound desc = could not find container \"8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832\": container with ID starting with 8ae15b1f7e216cc7fe8b653ed3fe12b5ca1f72cf10473b845f59a47597fa3832 not found: ID does not exist" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.517079 4809 scope.go:117] "RemoveContainer" containerID="1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.517434 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c"} err="failed to get container status \"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c\": rpc error: code = NotFound desc = could not find container \"1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c\": container with ID starting with 1358bdd3f8954b1467667e170e0eb747c54abfbedcacfc271cd4fc18f166027c not found: ID does not exist" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.520656 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.521040 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-metadata" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.521115 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-metadata" Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.521184 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad0cc4c-461f-4de2-aaf0-659c655edff3" containerName="nova-cell1-conductor-db-sync" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.521241 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad0cc4c-461f-4de2-aaf0-659c655edff3" containerName="nova-cell1-conductor-db-sync" Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.521301 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerName="init" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.521398 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerName="init" Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.521458 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-log" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.521515 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-log" Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.521572 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerName="dnsmasq-dns" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.521620 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerName="dnsmasq-dns" Dec 05 11:32:30 crc kubenswrapper[4809]: E1205 11:32:30.521713 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7d8ff5-5828-4f49-9400-bdb47aa56419" containerName="nova-manage" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.521764 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7d8ff5-5828-4f49-9400-bdb47aa56419" containerName="nova-manage" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.522038 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d52c2032-df32-4ff3-ab67-3c623e6fe80c" containerName="dnsmasq-dns" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.522115 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad0cc4c-461f-4de2-aaf0-659c655edff3" containerName="nova-cell1-conductor-db-sync" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.522177 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-log" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.522233 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7d8ff5-5828-4f49-9400-bdb47aa56419" containerName="nova-manage" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.522289 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" containerName="nova-metadata-metadata" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.522883 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.525841 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.552290 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.554370 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.558041 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.558076 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.573362 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.587525 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.641434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2wql\" (UniqueName: \"kubernetes.io/projected/cd1f3174-82d6-4264-8e70-01f2ea709403-kube-api-access-f2wql\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.641597 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.641667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.742878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c069cd4-4bde-4705-afa0-8330c0edd383-logs\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.742944 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2wql\" (UniqueName: \"kubernetes.io/projected/cd1f3174-82d6-4264-8e70-01f2ea709403-kube-api-access-f2wql\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.742993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-config-data\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.743034 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.743056 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.743075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.743113 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.743151 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dthrw\" (UniqueName: \"kubernetes.io/projected/4c069cd4-4bde-4705-afa0-8330c0edd383-kube-api-access-dthrw\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.748533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.752248 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.765523 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2wql\" (UniqueName: \"kubernetes.io/projected/cd1f3174-82d6-4264-8e70-01f2ea709403-kube-api-access-f2wql\") pod \"nova-cell1-conductor-0\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.844994 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c069cd4-4bde-4705-afa0-8330c0edd383-logs\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.845088 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-config-data\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.845159 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.845206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.845245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dthrw\" (UniqueName: \"kubernetes.io/projected/4c069cd4-4bde-4705-afa0-8330c0edd383-kube-api-access-dthrw\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.845757 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c069cd4-4bde-4705-afa0-8330c0edd383-logs\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.849872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-config-data\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.850779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.853282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.873330 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.873509 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dthrw\" (UniqueName: \"kubernetes.io/projected/4c069cd4-4bde-4705-afa0-8330c0edd383-kube-api-access-dthrw\") pod \"nova-metadata-0\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.887318 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:32:30 crc kubenswrapper[4809]: I1205 11:32:30.892449 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef450f73-b283-4a71-82ca-f93bbdfc31dc" path="/var/lib/kubelet/pods/ef450f73-b283-4a71-82ca-f93bbdfc31dc/volumes" Dec 05 11:32:31 crc kubenswrapper[4809]: W1205 11:32:31.355776 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd1f3174_82d6_4264_8e70_01f2ea709403.slice/crio-24a44518182efb547a6928bd33cd3f773d93467bbce99f6c8531966dee797c46 WatchSource:0}: Error finding container 24a44518182efb547a6928bd33cd3f773d93467bbce99f6c8531966dee797c46: Status 404 returned error can't find the container with id 24a44518182efb547a6928bd33cd3f773d93467bbce99f6c8531966dee797c46 Dec 05 11:32:31 crc kubenswrapper[4809]: I1205 11:32:31.358431 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:32:31 crc kubenswrapper[4809]: I1205 11:32:31.423460 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:32:31 crc kubenswrapper[4809]: I1205 11:32:31.472154 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c069cd4-4bde-4705-afa0-8330c0edd383","Type":"ContainerStarted","Data":"bc1ee4cfa87b4b8bbaf243ee9c3f72154fcd54d2fc9b5b547752ece32d61fe83"} Dec 05 11:32:31 crc kubenswrapper[4809]: I1205 11:32:31.477708 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cd1f3174-82d6-4264-8e70-01f2ea709403","Type":"ContainerStarted","Data":"24a44518182efb547a6928bd33cd3f773d93467bbce99f6c8531966dee797c46"} Dec 05 11:32:32 crc kubenswrapper[4809]: I1205 11:32:32.501431 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cd1f3174-82d6-4264-8e70-01f2ea709403","Type":"ContainerStarted","Data":"236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6"} Dec 05 11:32:32 crc kubenswrapper[4809]: I1205 11:32:32.501956 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:32 crc kubenswrapper[4809]: I1205 11:32:32.506359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c069cd4-4bde-4705-afa0-8330c0edd383","Type":"ContainerStarted","Data":"2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e"} Dec 05 11:32:32 crc kubenswrapper[4809]: I1205 11:32:32.506415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c069cd4-4bde-4705-afa0-8330c0edd383","Type":"ContainerStarted","Data":"187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa"} Dec 05 11:32:32 crc kubenswrapper[4809]: I1205 11:32:32.536112 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.53608691 podStartE2EDuration="2.53608691s" podCreationTimestamp="2025-12-05 11:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:32.52407174 +0000 UTC m=+1447.915048368" watchObservedRunningTime="2025-12-05 11:32:32.53608691 +0000 UTC m=+1447.927063508" Dec 05 11:32:32 crc kubenswrapper[4809]: I1205 11:32:32.559223 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.559202827 podStartE2EDuration="2.559202827s" podCreationTimestamp="2025-12-05 11:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:32.541847799 +0000 UTC m=+1447.932824437" watchObservedRunningTime="2025-12-05 11:32:32.559202827 +0000 UTC m=+1447.950179395" Dec 05 11:32:32 crc kubenswrapper[4809]: E1205 11:32:32.790121 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:32:32 crc kubenswrapper[4809]: E1205 11:32:32.792030 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:32:32 crc kubenswrapper[4809]: E1205 11:32:32.793531 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:32:32 crc kubenswrapper[4809]: E1205 11:32:32.793588 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="feccc379-3fb9-4271-a582-3412618bfd12" containerName="nova-scheduler-scheduler" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.357394 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.363285 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.531911 4809 generic.go:334] "Generic (PLEG): container finished" podID="feccc379-3fb9-4271-a582-3412618bfd12" containerID="0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" exitCode=0 Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.532022 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"feccc379-3fb9-4271-a582-3412618bfd12","Type":"ContainerDied","Data":"0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a"} Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.532063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"feccc379-3fb9-4271-a582-3412618bfd12","Type":"ContainerDied","Data":"50f848b9e33217ba3011da326b729463e22c8f854b8e7c86e533d6bbbe433406"} Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.532089 4809 scope.go:117] "RemoveContainer" containerID="0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.532302 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.535390 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-config-data\") pod \"89b7fd8e-9acc-4069-a652-b033f142cc6c\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.535439 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-combined-ca-bundle\") pod \"feccc379-3fb9-4271-a582-3412618bfd12\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.535666 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdq64\" (UniqueName: \"kubernetes.io/projected/feccc379-3fb9-4271-a582-3412618bfd12-kube-api-access-sdq64\") pod \"feccc379-3fb9-4271-a582-3412618bfd12\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.535759 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-config-data\") pod \"feccc379-3fb9-4271-a582-3412618bfd12\" (UID: \"feccc379-3fb9-4271-a582-3412618bfd12\") " Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.535839 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b7fd8e-9acc-4069-a652-b033f142cc6c-logs\") pod \"89b7fd8e-9acc-4069-a652-b033f142cc6c\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.535878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8s7r\" (UniqueName: \"kubernetes.io/projected/89b7fd8e-9acc-4069-a652-b033f142cc6c-kube-api-access-q8s7r\") pod \"89b7fd8e-9acc-4069-a652-b033f142cc6c\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.535905 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-combined-ca-bundle\") pod \"89b7fd8e-9acc-4069-a652-b033f142cc6c\" (UID: \"89b7fd8e-9acc-4069-a652-b033f142cc6c\") " Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.537797 4809 generic.go:334] "Generic (PLEG): container finished" podID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerID="ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59" exitCode=0 Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.537835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"89b7fd8e-9acc-4069-a652-b033f142cc6c","Type":"ContainerDied","Data":"ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59"} Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.537858 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"89b7fd8e-9acc-4069-a652-b033f142cc6c","Type":"ContainerDied","Data":"446aee61012549036f10012e22f2fd02e291178ea5aaa335150a423d74e8fab4"} Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.537869 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b7fd8e-9acc-4069-a652-b033f142cc6c-logs" (OuterVolumeSpecName: "logs") pod "89b7fd8e-9acc-4069-a652-b033f142cc6c" (UID: "89b7fd8e-9acc-4069-a652-b033f142cc6c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.537912 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.542026 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feccc379-3fb9-4271-a582-3412618bfd12-kube-api-access-sdq64" (OuterVolumeSpecName: "kube-api-access-sdq64") pod "feccc379-3fb9-4271-a582-3412618bfd12" (UID: "feccc379-3fb9-4271-a582-3412618bfd12"). InnerVolumeSpecName "kube-api-access-sdq64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.546388 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b7fd8e-9acc-4069-a652-b033f142cc6c-kube-api-access-q8s7r" (OuterVolumeSpecName: "kube-api-access-q8s7r") pod "89b7fd8e-9acc-4069-a652-b033f142cc6c" (UID: "89b7fd8e-9acc-4069-a652-b033f142cc6c"). InnerVolumeSpecName "kube-api-access-q8s7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.566431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "feccc379-3fb9-4271-a582-3412618bfd12" (UID: "feccc379-3fb9-4271-a582-3412618bfd12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.574239 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-config-data" (OuterVolumeSpecName: "config-data") pod "89b7fd8e-9acc-4069-a652-b033f142cc6c" (UID: "89b7fd8e-9acc-4069-a652-b033f142cc6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.574990 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-config-data" (OuterVolumeSpecName: "config-data") pod "feccc379-3fb9-4271-a582-3412618bfd12" (UID: "feccc379-3fb9-4271-a582-3412618bfd12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.583615 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89b7fd8e-9acc-4069-a652-b033f142cc6c" (UID: "89b7fd8e-9acc-4069-a652-b033f142cc6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.629252 4809 scope.go:117] "RemoveContainer" containerID="0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" Dec 05 11:32:34 crc kubenswrapper[4809]: E1205 11:32:34.629925 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a\": container with ID starting with 0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a not found: ID does not exist" containerID="0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.629965 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a"} err="failed to get container status \"0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a\": rpc error: code = NotFound desc = could not find container \"0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a\": container with ID starting with 0ba3c202de4277361155a6b7eca0191f9f324987ba6eae0b69e1b376ff79845a not found: ID does not exist" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.629995 4809 scope.go:117] "RemoveContainer" containerID="ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.637856 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.637890 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.637905 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdq64\" (UniqueName: \"kubernetes.io/projected/feccc379-3fb9-4271-a582-3412618bfd12-kube-api-access-sdq64\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.637918 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feccc379-3fb9-4271-a582-3412618bfd12-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.637929 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b7fd8e-9acc-4069-a652-b033f142cc6c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.637941 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8s7r\" (UniqueName: \"kubernetes.io/projected/89b7fd8e-9acc-4069-a652-b033f142cc6c-kube-api-access-q8s7r\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.637954 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b7fd8e-9acc-4069-a652-b033f142cc6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.652188 4809 scope.go:117] "RemoveContainer" containerID="f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.675059 4809 scope.go:117] "RemoveContainer" containerID="ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59" Dec 05 11:32:34 crc kubenswrapper[4809]: E1205 11:32:34.675471 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59\": container with ID starting with ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59 not found: ID does not exist" containerID="ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.675513 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59"} err="failed to get container status \"ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59\": rpc error: code = NotFound desc = could not find container \"ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59\": container with ID starting with ae69b460343cb2fd1888f38b2582e50b8deaed73c428cfcd330681fbfc6a9d59 not found: ID does not exist" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.675542 4809 scope.go:117] "RemoveContainer" containerID="f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6" Dec 05 11:32:34 crc kubenswrapper[4809]: E1205 11:32:34.675920 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6\": container with ID starting with f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6 not found: ID does not exist" containerID="f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.675952 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6"} err="failed to get container status \"f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6\": rpc error: code = NotFound desc = could not find container \"f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6\": container with ID starting with f31368d13180c4372b9157252327917150f95bcd4bd31611e5c52c10e16d8ab6 not found: ID does not exist" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.900972 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.901015 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.901035 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:34 crc kubenswrapper[4809]: E1205 11:32:34.906305 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-api" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.906542 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-api" Dec 05 11:32:34 crc kubenswrapper[4809]: E1205 11:32:34.906670 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feccc379-3fb9-4271-a582-3412618bfd12" containerName="nova-scheduler-scheduler" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.906728 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="feccc379-3fb9-4271-a582-3412618bfd12" containerName="nova-scheduler-scheduler" Dec 05 11:32:34 crc kubenswrapper[4809]: E1205 11:32:34.906819 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-log" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.906875 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-log" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.907677 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-api" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.907767 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" containerName="nova-api-log" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.907832 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="feccc379-3fb9-4271-a582-3412618bfd12" containerName="nova-scheduler-scheduler" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.910290 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.916701 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.974064 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:34 crc kubenswrapper[4809]: I1205 11:32:34.992312 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.001316 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.010847 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.012462 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.015138 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.035648 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.044539 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-config-data\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.044610 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdgfl\" (UniqueName: \"kubernetes.io/projected/18669644-e182-493a-8a14-2625f80e7e1f-kube-api-access-kdgfl\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.044878 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.147522 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhpxb\" (UniqueName: \"kubernetes.io/projected/93e4202c-630d-47a7-a6a9-c0259a45ec58-kube-api-access-jhpxb\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.147653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-config-data\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.147744 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdgfl\" (UniqueName: \"kubernetes.io/projected/18669644-e182-493a-8a14-2625f80e7e1f-kube-api-access-kdgfl\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.147846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e4202c-630d-47a7-a6a9-c0259a45ec58-logs\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.147987 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.148039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.148178 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-config-data\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.154466 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-config-data\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.155726 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.181017 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdgfl\" (UniqueName: \"kubernetes.io/projected/18669644-e182-493a-8a14-2625f80e7e1f-kube-api-access-kdgfl\") pod \"nova-scheduler-0\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.250073 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-config-data\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.250223 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhpxb\" (UniqueName: \"kubernetes.io/projected/93e4202c-630d-47a7-a6a9-c0259a45ec58-kube-api-access-jhpxb\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.250274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e4202c-630d-47a7-a6a9-c0259a45ec58-logs\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.250321 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.250614 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e4202c-630d-47a7-a6a9-c0259a45ec58-logs\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.253405 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-config-data\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.262458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.265421 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhpxb\" (UniqueName: \"kubernetes.io/projected/93e4202c-630d-47a7-a6a9-c0259a45ec58-kube-api-access-jhpxb\") pod \"nova-api-0\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.273699 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.332661 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.794018 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.843002 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:35 crc kubenswrapper[4809]: W1205 11:32:35.873779 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e4202c_630d_47a7_a6a9_c0259a45ec58.slice/crio-078ee16637e453aea149cc85463b15d27780a35622d417f4d0afe12eb2e88106 WatchSource:0}: Error finding container 078ee16637e453aea149cc85463b15d27780a35622d417f4d0afe12eb2e88106: Status 404 returned error can't find the container with id 078ee16637e453aea149cc85463b15d27780a35622d417f4d0afe12eb2e88106 Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.888025 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:32:35 crc kubenswrapper[4809]: I1205 11:32:35.888066 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.563218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18669644-e182-493a-8a14-2625f80e7e1f","Type":"ContainerStarted","Data":"65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d"} Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.563539 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18669644-e182-493a-8a14-2625f80e7e1f","Type":"ContainerStarted","Data":"eb93806dc4c131e3bc3a4e5eb5eba24f58e821d3a036dbae82fd76aaea8f5a15"} Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.565449 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93e4202c-630d-47a7-a6a9-c0259a45ec58","Type":"ContainerStarted","Data":"6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604"} Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.565498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93e4202c-630d-47a7-a6a9-c0259a45ec58","Type":"ContainerStarted","Data":"ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958"} Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.565512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93e4202c-630d-47a7-a6a9-c0259a45ec58","Type":"ContainerStarted","Data":"078ee16637e453aea149cc85463b15d27780a35622d417f4d0afe12eb2e88106"} Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.595397 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5953778610000002 podStartE2EDuration="2.595377861s" podCreationTimestamp="2025-12-05 11:32:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:36.583161115 +0000 UTC m=+1451.974137693" watchObservedRunningTime="2025-12-05 11:32:36.595377861 +0000 UTC m=+1451.986354429" Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.607258 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.607233977 podStartE2EDuration="2.607233977s" podCreationTimestamp="2025-12-05 11:32:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:36.603255824 +0000 UTC m=+1451.994232422" watchObservedRunningTime="2025-12-05 11:32:36.607233977 +0000 UTC m=+1451.998210545" Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.886800 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b7fd8e-9acc-4069-a652-b033f142cc6c" path="/var/lib/kubelet/pods/89b7fd8e-9acc-4069-a652-b033f142cc6c/volumes" Dec 05 11:32:36 crc kubenswrapper[4809]: I1205 11:32:36.887559 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feccc379-3fb9-4271-a582-3412618bfd12" path="/var/lib/kubelet/pods/feccc379-3fb9-4271-a582-3412618bfd12/volumes" Dec 05 11:32:40 crc kubenswrapper[4809]: I1205 11:32:40.277148 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 11:32:40 crc kubenswrapper[4809]: I1205 11:32:40.888675 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:32:40 crc kubenswrapper[4809]: I1205 11:32:40.888724 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:32:40 crc kubenswrapper[4809]: I1205 11:32:40.918360 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 11:32:41 crc kubenswrapper[4809]: I1205 11:32:41.902832 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:32:41 crc kubenswrapper[4809]: I1205 11:32:41.902863 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:32:45 crc kubenswrapper[4809]: I1205 11:32:45.274603 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 11:32:45 crc kubenswrapper[4809]: I1205 11:32:45.331790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 11:32:45 crc kubenswrapper[4809]: I1205 11:32:45.334432 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:32:45 crc kubenswrapper[4809]: I1205 11:32:45.334508 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:32:45 crc kubenswrapper[4809]: I1205 11:32:45.674250 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 11:32:46 crc kubenswrapper[4809]: I1205 11:32:46.417882 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:32:46 crc kubenswrapper[4809]: I1205 11:32:46.417887 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:32:50 crc kubenswrapper[4809]: I1205 11:32:50.897821 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 11:32:50 crc kubenswrapper[4809]: I1205 11:32:50.899065 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 11:32:50 crc kubenswrapper[4809]: I1205 11:32:50.903179 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 11:32:51 crc kubenswrapper[4809]: I1205 11:32:51.723528 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.728618 4809 generic.go:334] "Generic (PLEG): container finished" podID="15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" containerID="894c6277b8486291c59a23e96d6969fc19f092834afc5ce802d6b8dc9d69bcc6" exitCode=137 Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.728665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf","Type":"ContainerDied","Data":"894c6277b8486291c59a23e96d6969fc19f092834afc5ce802d6b8dc9d69bcc6"} Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.728736 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf","Type":"ContainerDied","Data":"afb10f41a5a60585b85d154b55a42fc59b750d7d1adb3cd93fafbdfb077dea25"} Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.728754 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afb10f41a5a60585b85d154b55a42fc59b750d7d1adb3cd93fafbdfb077dea25" Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.757871 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.931786 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9x9w\" (UniqueName: \"kubernetes.io/projected/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-kube-api-access-r9x9w\") pod \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.931901 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-combined-ca-bundle\") pod \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.931929 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-config-data\") pod \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\" (UID: \"15adc9f9-f77a-4ef9-bd25-e7de9fac53bf\") " Dec 05 11:32:52 crc kubenswrapper[4809]: I1205 11:32:52.952830 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-kube-api-access-r9x9w" (OuterVolumeSpecName: "kube-api-access-r9x9w") pod "15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" (UID: "15adc9f9-f77a-4ef9-bd25-e7de9fac53bf"). InnerVolumeSpecName "kube-api-access-r9x9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.008203 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" (UID: "15adc9f9-f77a-4ef9-bd25-e7de9fac53bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.009540 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-config-data" (OuterVolumeSpecName: "config-data") pod "15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" (UID: "15adc9f9-f77a-4ef9-bd25-e7de9fac53bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.034466 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9x9w\" (UniqueName: \"kubernetes.io/projected/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-kube-api-access-r9x9w\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.034502 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.034512 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.739792 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.782943 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.800149 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.816836 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:53 crc kubenswrapper[4809]: E1205 11:32:53.818543 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.818660 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.818977 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.819957 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.824164 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.824467 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.828960 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.834471 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.954020 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.954106 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.954172 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.954499 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:53 crc kubenswrapper[4809]: I1205 11:32:53.954847 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf8cr\" (UniqueName: \"kubernetes.io/projected/50ec8106-d71a-45e3-af2d-f9bf27061233-kube-api-access-mf8cr\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.057092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.057275 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.057468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf8cr\" (UniqueName: \"kubernetes.io/projected/50ec8106-d71a-45e3-af2d-f9bf27061233-kube-api-access-mf8cr\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.057617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.057774 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.063927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.063944 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.065028 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.065066 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.077409 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf8cr\" (UniqueName: \"kubernetes.io/projected/50ec8106-d71a-45e3-af2d-f9bf27061233-kube-api-access-mf8cr\") pod \"nova-cell1-novncproxy-0\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.142152 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.838597 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:32:54 crc kubenswrapper[4809]: W1205 11:32:54.845742 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50ec8106_d71a_45e3_af2d_f9bf27061233.slice/crio-212abebcd84168a4904d4dc0a7ee58aae12623d56536dec5af88f36cd1afa7d0 WatchSource:0}: Error finding container 212abebcd84168a4904d4dc0a7ee58aae12623d56536dec5af88f36cd1afa7d0: Status 404 returned error can't find the container with id 212abebcd84168a4904d4dc0a7ee58aae12623d56536dec5af88f36cd1afa7d0 Dec 05 11:32:54 crc kubenswrapper[4809]: I1205 11:32:54.892595 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15adc9f9-f77a-4ef9-bd25-e7de9fac53bf" path="/var/lib/kubelet/pods/15adc9f9-f77a-4ef9-bd25-e7de9fac53bf/volumes" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.338112 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.338750 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.341772 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.343325 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.787572 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50ec8106-d71a-45e3-af2d-f9bf27061233","Type":"ContainerStarted","Data":"4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f"} Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.787648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50ec8106-d71a-45e3-af2d-f9bf27061233","Type":"ContainerStarted","Data":"212abebcd84168a4904d4dc0a7ee58aae12623d56536dec5af88f36cd1afa7d0"} Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.787939 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.802107 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.816919 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.816897198 podStartE2EDuration="2.816897198s" podCreationTimestamp="2025-12-05 11:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:55.805666278 +0000 UTC m=+1471.196642836" watchObservedRunningTime="2025-12-05 11:32:55.816897198 +0000 UTC m=+1471.207873766" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.976719 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ch8g6"] Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.979100 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:55 crc kubenswrapper[4809]: I1205 11:32:55.996944 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ch8g6"] Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.099271 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.099363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-config\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.099553 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.099694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.099742 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.099827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66227\" (UniqueName: \"kubernetes.io/projected/462b7e18-c84b-42e8-ad9f-379bfe01bd26-kube-api-access-66227\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.201861 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.202134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.202187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66227\" (UniqueName: \"kubernetes.io/projected/462b7e18-c84b-42e8-ad9f-379bfe01bd26-kube-api-access-66227\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.202244 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.202316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-config\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.202354 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.203376 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.203471 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.203512 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.203479 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-config\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.203934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.231424 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66227\" (UniqueName: \"kubernetes.io/projected/462b7e18-c84b-42e8-ad9f-379bfe01bd26-kube-api-access-66227\") pod \"dnsmasq-dns-59cf4bdb65-ch8g6\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.296793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:56 crc kubenswrapper[4809]: I1205 11:32:56.808168 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ch8g6"] Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.807919 4809 generic.go:334] "Generic (PLEG): container finished" podID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerID="ceba1f0259ef6c16939487385f592f769bf0fd889b7aabff233d7301f28ef280" exitCode=0 Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.809198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" event={"ID":"462b7e18-c84b-42e8-ad9f-379bfe01bd26","Type":"ContainerDied","Data":"ceba1f0259ef6c16939487385f592f769bf0fd889b7aabff233d7301f28ef280"} Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.809233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" event={"ID":"462b7e18-c84b-42e8-ad9f-379bfe01bd26","Type":"ContainerStarted","Data":"1656539d6fa79db56b38220e98941c55ca11a0efd1df637efdbbac04b8c6c519"} Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.918280 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.918896 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-central-agent" containerID="cri-o://7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e" gracePeriod=30 Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.919370 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-notification-agent" containerID="cri-o://8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c" gracePeriod=30 Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.919494 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="sg-core" containerID="cri-o://c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a" gracePeriod=30 Dec 05 11:32:57 crc kubenswrapper[4809]: I1205 11:32:57.924738 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="proxy-httpd" containerID="cri-o://3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9" gracePeriod=30 Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.826714 4809 generic.go:334] "Generic (PLEG): container finished" podID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerID="3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9" exitCode=0 Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.827035 4809 generic.go:334] "Generic (PLEG): container finished" podID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerID="c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a" exitCode=2 Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.827043 4809 generic.go:334] "Generic (PLEG): container finished" podID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerID="7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e" exitCode=0 Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.826784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerDied","Data":"3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9"} Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.827143 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerDied","Data":"c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a"} Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.827179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerDied","Data":"7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e"} Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.830053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" event={"ID":"462b7e18-c84b-42e8-ad9f-379bfe01bd26","Type":"ContainerStarted","Data":"bd9495b849c3fa62b2b01b06cf243c2210d8b84430e1a712ed2a4c411f4baa90"} Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.830289 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.864295 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" podStartSLOduration=3.864275381 podStartE2EDuration="3.864275381s" podCreationTimestamp="2025-12-05 11:32:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:32:58.854313724 +0000 UTC m=+1474.245290282" watchObservedRunningTime="2025-12-05 11:32:58.864275381 +0000 UTC m=+1474.255251939" Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.993292 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.993542 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-log" containerID="cri-o://ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958" gracePeriod=30 Dec 05 11:32:58 crc kubenswrapper[4809]: I1205 11:32:58.993667 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-api" containerID="cri-o://6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604" gracePeriod=30 Dec 05 11:32:59 crc kubenswrapper[4809]: I1205 11:32:59.142910 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:32:59 crc kubenswrapper[4809]: I1205 11:32:59.840369 4809 generic.go:334] "Generic (PLEG): container finished" podID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerID="ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958" exitCode=143 Dec 05 11:32:59 crc kubenswrapper[4809]: I1205 11:32:59.840454 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93e4202c-630d-47a7-a6a9-c0259a45ec58","Type":"ContainerDied","Data":"ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958"} Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.558304 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.685348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-combined-ca-bundle\") pod \"93e4202c-630d-47a7-a6a9-c0259a45ec58\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.685484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-config-data\") pod \"93e4202c-630d-47a7-a6a9-c0259a45ec58\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.685672 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e4202c-630d-47a7-a6a9-c0259a45ec58-logs\") pod \"93e4202c-630d-47a7-a6a9-c0259a45ec58\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.686106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93e4202c-630d-47a7-a6a9-c0259a45ec58-logs" (OuterVolumeSpecName: "logs") pod "93e4202c-630d-47a7-a6a9-c0259a45ec58" (UID: "93e4202c-630d-47a7-a6a9-c0259a45ec58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.686264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhpxb\" (UniqueName: \"kubernetes.io/projected/93e4202c-630d-47a7-a6a9-c0259a45ec58-kube-api-access-jhpxb\") pod \"93e4202c-630d-47a7-a6a9-c0259a45ec58\" (UID: \"93e4202c-630d-47a7-a6a9-c0259a45ec58\") " Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.686733 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e4202c-630d-47a7-a6a9-c0259a45ec58-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.708949 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e4202c-630d-47a7-a6a9-c0259a45ec58-kube-api-access-jhpxb" (OuterVolumeSpecName: "kube-api-access-jhpxb") pod "93e4202c-630d-47a7-a6a9-c0259a45ec58" (UID: "93e4202c-630d-47a7-a6a9-c0259a45ec58"). InnerVolumeSpecName "kube-api-access-jhpxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.717624 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93e4202c-630d-47a7-a6a9-c0259a45ec58" (UID: "93e4202c-630d-47a7-a6a9-c0259a45ec58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.720840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-config-data" (OuterVolumeSpecName: "config-data") pod "93e4202c-630d-47a7-a6a9-c0259a45ec58" (UID: "93e4202c-630d-47a7-a6a9-c0259a45ec58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.787953 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.787995 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e4202c-630d-47a7-a6a9-c0259a45ec58-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.788008 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhpxb\" (UniqueName: \"kubernetes.io/projected/93e4202c-630d-47a7-a6a9-c0259a45ec58-kube-api-access-jhpxb\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.872483 4809 generic.go:334] "Generic (PLEG): container finished" podID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerID="6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604" exitCode=0 Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.872535 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93e4202c-630d-47a7-a6a9-c0259a45ec58","Type":"ContainerDied","Data":"6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604"} Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.872565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93e4202c-630d-47a7-a6a9-c0259a45ec58","Type":"ContainerDied","Data":"078ee16637e453aea149cc85463b15d27780a35622d417f4d0afe12eb2e88106"} Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.872590 4809 scope.go:117] "RemoveContainer" containerID="6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.872853 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.902335 4809 scope.go:117] "RemoveContainer" containerID="ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.926084 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.931836 4809 scope.go:117] "RemoveContainer" containerID="6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604" Dec 05 11:33:02 crc kubenswrapper[4809]: E1205 11:33:02.932871 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604\": container with ID starting with 6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604 not found: ID does not exist" containerID="6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.932920 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604"} err="failed to get container status \"6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604\": rpc error: code = NotFound desc = could not find container \"6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604\": container with ID starting with 6bd851a7386662c113ef5f6607c553edb77b3466ac8126dc89362edde6a39604 not found: ID does not exist" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.932949 4809 scope.go:117] "RemoveContainer" containerID="ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958" Dec 05 11:33:02 crc kubenswrapper[4809]: E1205 11:33:02.933434 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958\": container with ID starting with ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958 not found: ID does not exist" containerID="ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.933476 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958"} err="failed to get container status \"ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958\": rpc error: code = NotFound desc = could not find container \"ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958\": container with ID starting with ecdac0929f67da8220d0b77ccae7f78d0b0b07c81561a65d05c15c1af2bd1958 not found: ID does not exist" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.940121 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.952123 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:02 crc kubenswrapper[4809]: E1205 11:33:02.952713 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-log" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.952746 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-log" Dec 05 11:33:02 crc kubenswrapper[4809]: E1205 11:33:02.952810 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-api" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.952822 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-api" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.953155 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-api" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.953206 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" containerName="nova-api-log" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.955039 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.960321 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.961605 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.966876 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 11:33:02 crc kubenswrapper[4809]: I1205 11:33:02.967040 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.095340 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5snhb\" (UniqueName: \"kubernetes.io/projected/83a4702d-662e-4c2b-8335-d5f43099c44b-kube-api-access-5snhb\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.095403 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.095552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-public-tls-certs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.095690 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.095725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83a4702d-662e-4c2b-8335-d5f43099c44b-logs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.095770 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-config-data\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.197206 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.197488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83a4702d-662e-4c2b-8335-d5f43099c44b-logs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.197542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-config-data\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.197661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5snhb\" (UniqueName: \"kubernetes.io/projected/83a4702d-662e-4c2b-8335-d5f43099c44b-kube-api-access-5snhb\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.197693 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.197729 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-public-tls-certs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.198122 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83a4702d-662e-4c2b-8335-d5f43099c44b-logs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.209540 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-public-tls-certs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.209569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.210178 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.211289 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-config-data\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.215827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5snhb\" (UniqueName: \"kubernetes.io/projected/83a4702d-662e-4c2b-8335-d5f43099c44b-kube-api-access-5snhb\") pod \"nova-api-0\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.279098 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.757950 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:03 crc kubenswrapper[4809]: I1205 11:33:03.889670 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83a4702d-662e-4c2b-8335-d5f43099c44b","Type":"ContainerStarted","Data":"67b3fe0142c80e79a10796ea7321b89266b3f8a6bdfbd2e8e8ec2e4edea4fc6a"} Dec 05 11:33:04 crc kubenswrapper[4809]: I1205 11:33:04.142866 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:33:04 crc kubenswrapper[4809]: I1205 11:33:04.158163 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:33:04 crc kubenswrapper[4809]: I1205 11:33:04.899584 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e4202c-630d-47a7-a6a9-c0259a45ec58" path="/var/lib/kubelet/pods/93e4202c-630d-47a7-a6a9-c0259a45ec58/volumes" Dec 05 11:33:04 crc kubenswrapper[4809]: I1205 11:33:04.912368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83a4702d-662e-4c2b-8335-d5f43099c44b","Type":"ContainerStarted","Data":"99c7a952c43dee727c2d091290d87f6671731e8e5972b24248e47525e3da6a3a"} Dec 05 11:33:04 crc kubenswrapper[4809]: I1205 11:33:04.912420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83a4702d-662e-4c2b-8335-d5f43099c44b","Type":"ContainerStarted","Data":"fc8f6fbd0d6af548b3ab28ac3220e10e5ec059642c2119b7cb235f5f82975b77"} Dec 05 11:33:04 crc kubenswrapper[4809]: I1205 11:33:04.939820 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.939799621 podStartE2EDuration="2.939799621s" podCreationTimestamp="2025-12-05 11:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:33:04.929612638 +0000 UTC m=+1480.320589216" watchObservedRunningTime="2025-12-05 11:33:04.939799621 +0000 UTC m=+1480.330776179" Dec 05 11:33:04 crc kubenswrapper[4809]: I1205 11:33:04.943347 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.123322 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-qnwjc"] Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.124800 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.134377 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.135116 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.146728 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-qnwjc"] Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.255467 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nvzx\" (UniqueName: \"kubernetes.io/projected/cae5c530-0398-48a0-ad06-2b08fd0d9c44-kube-api-access-2nvzx\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.255864 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.255976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-scripts\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.256052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-config-data\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.357976 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.358300 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-scripts\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.358329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-config-data\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.358412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nvzx\" (UniqueName: \"kubernetes.io/projected/cae5c530-0398-48a0-ad06-2b08fd0d9c44-kube-api-access-2nvzx\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.362971 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-scripts\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.363032 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.363439 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-config-data\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.375225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nvzx\" (UniqueName: \"kubernetes.io/projected/cae5c530-0398-48a0-ad06-2b08fd0d9c44-kube-api-access-2nvzx\") pod \"nova-cell1-cell-mapping-qnwjc\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.451186 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:05 crc kubenswrapper[4809]: I1205 11:33:05.934374 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-qnwjc"] Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.297830 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.379401 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-2pv8s"] Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.379722 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" podUID="af9a1e65-a606-4221-96e2-72e11675fd31" containerName="dnsmasq-dns" containerID="cri-o://7263bb11bc485b26900d63cf7d94328801219440e7d4e00a3937691228bd52c7" gracePeriod=10 Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.940974 4809 generic.go:334] "Generic (PLEG): container finished" podID="af9a1e65-a606-4221-96e2-72e11675fd31" containerID="7263bb11bc485b26900d63cf7d94328801219440e7d4e00a3937691228bd52c7" exitCode=0 Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.941352 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" event={"ID":"af9a1e65-a606-4221-96e2-72e11675fd31","Type":"ContainerDied","Data":"7263bb11bc485b26900d63cf7d94328801219440e7d4e00a3937691228bd52c7"} Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.943821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qnwjc" event={"ID":"cae5c530-0398-48a0-ad06-2b08fd0d9c44","Type":"ContainerStarted","Data":"cee2aebc7c4c395069fa4d92acc3a5a78a920e62eb51e07edd4caaa0df65aaca"} Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.943854 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qnwjc" event={"ID":"cae5c530-0398-48a0-ad06-2b08fd0d9c44","Type":"ContainerStarted","Data":"d085dd685d14158532bd2442d264320133a43b843ed67c8bb19c4314bbd03459"} Dec 05 11:33:06 crc kubenswrapper[4809]: I1205 11:33:06.964527 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-qnwjc" podStartSLOduration=1.964505268 podStartE2EDuration="1.964505268s" podCreationTimestamp="2025-12-05 11:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:33:06.958663997 +0000 UTC m=+1482.349640575" watchObservedRunningTime="2025-12-05 11:33:06.964505268 +0000 UTC m=+1482.355481836" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.023414 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.028303 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-config\") pod \"af9a1e65-a606-4221-96e2-72e11675fd31\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.028346 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-svc\") pod \"af9a1e65-a606-4221-96e2-72e11675fd31\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.028421 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-nb\") pod \"af9a1e65-a606-4221-96e2-72e11675fd31\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.028446 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-swift-storage-0\") pod \"af9a1e65-a606-4221-96e2-72e11675fd31\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.119530 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "af9a1e65-a606-4221-96e2-72e11675fd31" (UID: "af9a1e65-a606-4221-96e2-72e11675fd31"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.130473 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvlrg\" (UniqueName: \"kubernetes.io/projected/af9a1e65-a606-4221-96e2-72e11675fd31-kube-api-access-cvlrg\") pod \"af9a1e65-a606-4221-96e2-72e11675fd31\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.130669 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-sb\") pod \"af9a1e65-a606-4221-96e2-72e11675fd31\" (UID: \"af9a1e65-a606-4221-96e2-72e11675fd31\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.131530 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.133616 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9a1e65-a606-4221-96e2-72e11675fd31-kube-api-access-cvlrg" (OuterVolumeSpecName: "kube-api-access-cvlrg") pod "af9a1e65-a606-4221-96e2-72e11675fd31" (UID: "af9a1e65-a606-4221-96e2-72e11675fd31"). InnerVolumeSpecName "kube-api-access-cvlrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.144551 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-config" (OuterVolumeSpecName: "config") pod "af9a1e65-a606-4221-96e2-72e11675fd31" (UID: "af9a1e65-a606-4221-96e2-72e11675fd31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.154550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "af9a1e65-a606-4221-96e2-72e11675fd31" (UID: "af9a1e65-a606-4221-96e2-72e11675fd31"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.165439 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "af9a1e65-a606-4221-96e2-72e11675fd31" (UID: "af9a1e65-a606-4221-96e2-72e11675fd31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.193643 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "af9a1e65-a606-4221-96e2-72e11675fd31" (UID: "af9a1e65-a606-4221-96e2-72e11675fd31"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.233937 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.233971 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.233985 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.233999 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvlrg\" (UniqueName: \"kubernetes.io/projected/af9a1e65-a606-4221-96e2-72e11675fd31-kube-api-access-cvlrg\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.234010 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af9a1e65-a606-4221-96e2-72e11675fd31-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.470089 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.640530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-ceilometer-tls-certs\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.640681 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-scripts\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.640739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqkj4\" (UniqueName: \"kubernetes.io/projected/139b3658-4df4-4cff-9a73-1e20624b8d8f-kube-api-access-dqkj4\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.640803 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-combined-ca-bundle\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.640859 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-config-data\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.640891 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-run-httpd\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.640973 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-sg-core-conf-yaml\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.641004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-log-httpd\") pod \"139b3658-4df4-4cff-9a73-1e20624b8d8f\" (UID: \"139b3658-4df4-4cff-9a73-1e20624b8d8f\") " Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.641788 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.641884 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.659945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-scripts" (OuterVolumeSpecName: "scripts") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.661909 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/139b3658-4df4-4cff-9a73-1e20624b8d8f-kube-api-access-dqkj4" (OuterVolumeSpecName: "kube-api-access-dqkj4") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "kube-api-access-dqkj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.674810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.714356 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.743488 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.743518 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.743528 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqkj4\" (UniqueName: \"kubernetes.io/projected/139b3658-4df4-4cff-9a73-1e20624b8d8f-kube-api-access-dqkj4\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.743539 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.743547 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.743556 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/139b3658-4df4-4cff-9a73-1e20624b8d8f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.751256 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.761207 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-config-data" (OuterVolumeSpecName: "config-data") pod "139b3658-4df4-4cff-9a73-1e20624b8d8f" (UID: "139b3658-4df4-4cff-9a73-1e20624b8d8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.845744 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.845791 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139b3658-4df4-4cff-9a73-1e20624b8d8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.958951 4809 generic.go:334] "Generic (PLEG): container finished" podID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerID="8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c" exitCode=0 Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.959021 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerDied","Data":"8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c"} Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.959024 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.959049 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"139b3658-4df4-4cff-9a73-1e20624b8d8f","Type":"ContainerDied","Data":"9d2e380550da32aeeabc3b3c3426824248fded444568d9aa18705297e7481a6a"} Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.959069 4809 scope.go:117] "RemoveContainer" containerID="3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.965946 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.966052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-2pv8s" event={"ID":"af9a1e65-a606-4221-96e2-72e11675fd31","Type":"ContainerDied","Data":"2b22c1f28c422cc5d8682b3df064105ef903bde29290ffb106a21fef064706d5"} Dec 05 11:33:07 crc kubenswrapper[4809]: I1205 11:33:07.990769 4809 scope.go:117] "RemoveContainer" containerID="c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.016509 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.025266 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.029802 4809 scope.go:117] "RemoveContainer" containerID="8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.044131 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-2pv8s"] Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.061390 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-2pv8s"] Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.068819 4809 scope.go:117] "RemoveContainer" containerID="7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.073587 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.074125 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9a1e65-a606-4221-96e2-72e11675fd31" containerName="init" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074159 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9a1e65-a606-4221-96e2-72e11675fd31" containerName="init" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.074172 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-notification-agent" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074180 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-notification-agent" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.074190 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="sg-core" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074198 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="sg-core" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.074235 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9a1e65-a606-4221-96e2-72e11675fd31" containerName="dnsmasq-dns" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074244 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9a1e65-a606-4221-96e2-72e11675fd31" containerName="dnsmasq-dns" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.074260 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-central-agent" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074267 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-central-agent" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.074285 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="proxy-httpd" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074292 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="proxy-httpd" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074504 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9a1e65-a606-4221-96e2-72e11675fd31" containerName="dnsmasq-dns" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074521 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="proxy-httpd" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074546 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="sg-core" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074566 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-central-agent" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.074582 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" containerName="ceilometer-notification-agent" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.077197 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.081039 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.082230 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.082685 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.083237 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.111440 4809 scope.go:117] "RemoveContainer" containerID="3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.111945 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9\": container with ID starting with 3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9 not found: ID does not exist" containerID="3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.111985 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9"} err="failed to get container status \"3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9\": rpc error: code = NotFound desc = could not find container \"3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9\": container with ID starting with 3f198fb3059c4991434bea042bc04a89935466854daeb79b2ba41ebbe4c2e3e9 not found: ID does not exist" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.112011 4809 scope.go:117] "RemoveContainer" containerID="c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.112839 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a\": container with ID starting with c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a not found: ID does not exist" containerID="c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.112870 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a"} err="failed to get container status \"c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a\": rpc error: code = NotFound desc = could not find container \"c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a\": container with ID starting with c3986a3b0e3c9515fb92ffcba21caf24932934d945ae853a3c0ca19450affb4a not found: ID does not exist" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.112889 4809 scope.go:117] "RemoveContainer" containerID="8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.113321 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c\": container with ID starting with 8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c not found: ID does not exist" containerID="8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.113375 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c"} err="failed to get container status \"8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c\": rpc error: code = NotFound desc = could not find container \"8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c\": container with ID starting with 8a04acf6a64e3c13cd28103c1908aac899bce78e0e21404e6a20cf2e8da4974c not found: ID does not exist" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.113393 4809 scope.go:117] "RemoveContainer" containerID="7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.113769 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e\": container with ID starting with 7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e not found: ID does not exist" containerID="7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.113803 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e"} err="failed to get container status \"7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e\": rpc error: code = NotFound desc = could not find container \"7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e\": container with ID starting with 7815eac361002ad6416aa20300cbded070def4df4c3c6d9c92dc6df10573862e not found: ID does not exist" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.113816 4809 scope.go:117] "RemoveContainer" containerID="7263bb11bc485b26900d63cf7d94328801219440e7d4e00a3937691228bd52c7" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.132899 4809 scope.go:117] "RemoveContainer" containerID="b6e1b17faedc5097d9b2d6c223245dc1c85527026a411fd27c7c506de3918ad2" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.151742 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-scripts\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.152058 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-log-httpd\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.152151 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.152230 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmbbv\" (UniqueName: \"kubernetes.io/projected/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-kube-api-access-kmbbv\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.152320 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.152462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-run-httpd\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.152571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-config-data\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.152656 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: E1205 11:33:08.161938 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod139b3658_4df4_4cff_9a73_1e20624b8d8f.slice/crio-9d2e380550da32aeeabc3b3c3426824248fded444568d9aa18705297e7481a6a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod139b3658_4df4_4cff_9a73_1e20624b8d8f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf9a1e65_a606_4221_96e2_72e11675fd31.slice/crio-2b22c1f28c422cc5d8682b3df064105ef903bde29290ffb106a21fef064706d5\": RecentStats: unable to find data in memory cache]" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.254238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-run-httpd\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255083 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-run-httpd\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255156 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-config-data\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255192 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255233 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-scripts\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-log-httpd\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255355 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255381 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmbbv\" (UniqueName: \"kubernetes.io/projected/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-kube-api-access-kmbbv\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.255415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.256753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-log-httpd\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.261154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.262554 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-config-data\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.263013 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-scripts\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.263174 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.272489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.275912 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmbbv\" (UniqueName: \"kubernetes.io/projected/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-kube-api-access-kmbbv\") pod \"ceilometer-0\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " pod="openstack/ceilometer-0" Dec 05 11:33:08 crc kubenswrapper[4809]: I1205 11:33:08.411731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:33:09 crc kubenswrapper[4809]: I1205 11:33:08.894392 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="139b3658-4df4-4cff-9a73-1e20624b8d8f" path="/var/lib/kubelet/pods/139b3658-4df4-4cff-9a73-1e20624b8d8f/volumes" Dec 05 11:33:09 crc kubenswrapper[4809]: I1205 11:33:08.897620 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9a1e65-a606-4221-96e2-72e11675fd31" path="/var/lib/kubelet/pods/af9a1e65-a606-4221-96e2-72e11675fd31/volumes" Dec 05 11:33:09 crc kubenswrapper[4809]: I1205 11:33:08.904064 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:33:09 crc kubenswrapper[4809]: I1205 11:33:08.907613 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:33:09 crc kubenswrapper[4809]: I1205 11:33:08.974461 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerStarted","Data":"ecc175d8ec4a837fb42632f9662bf5c71bc6bbb60432c51bf1729918933c1c30"} Dec 05 11:33:09 crc kubenswrapper[4809]: I1205 11:33:09.987924 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerStarted","Data":"6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898"} Dec 05 11:33:11 crc kubenswrapper[4809]: I1205 11:33:11.017537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerStarted","Data":"8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458"} Dec 05 11:33:12 crc kubenswrapper[4809]: I1205 11:33:12.032891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerStarted","Data":"d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509"} Dec 05 11:33:12 crc kubenswrapper[4809]: I1205 11:33:12.039776 4809 generic.go:334] "Generic (PLEG): container finished" podID="cae5c530-0398-48a0-ad06-2b08fd0d9c44" containerID="cee2aebc7c4c395069fa4d92acc3a5a78a920e62eb51e07edd4caaa0df65aaca" exitCode=0 Dec 05 11:33:12 crc kubenswrapper[4809]: I1205 11:33:12.039823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qnwjc" event={"ID":"cae5c530-0398-48a0-ad06-2b08fd0d9c44","Type":"ContainerDied","Data":"cee2aebc7c4c395069fa4d92acc3a5a78a920e62eb51e07edd4caaa0df65aaca"} Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.054196 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerStarted","Data":"36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292"} Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.054840 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.086246 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.2703579 podStartE2EDuration="5.08621741s" podCreationTimestamp="2025-12-05 11:33:08 +0000 UTC" firstStartedPulling="2025-12-05 11:33:08.9037032 +0000 UTC m=+1484.294679758" lastFinishedPulling="2025-12-05 11:33:12.71956269 +0000 UTC m=+1488.110539268" observedRunningTime="2025-12-05 11:33:13.074685323 +0000 UTC m=+1488.465661901" watchObservedRunningTime="2025-12-05 11:33:13.08621741 +0000 UTC m=+1488.477193968" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.279887 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.280099 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.449182 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.468943 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-combined-ca-bundle\") pod \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.469143 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nvzx\" (UniqueName: \"kubernetes.io/projected/cae5c530-0398-48a0-ad06-2b08fd0d9c44-kube-api-access-2nvzx\") pod \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.469263 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-scripts\") pod \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.469307 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-config-data\") pod \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\" (UID: \"cae5c530-0398-48a0-ad06-2b08fd0d9c44\") " Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.475050 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-scripts" (OuterVolumeSpecName: "scripts") pod "cae5c530-0398-48a0-ad06-2b08fd0d9c44" (UID: "cae5c530-0398-48a0-ad06-2b08fd0d9c44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.475946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae5c530-0398-48a0-ad06-2b08fd0d9c44-kube-api-access-2nvzx" (OuterVolumeSpecName: "kube-api-access-2nvzx") pod "cae5c530-0398-48a0-ad06-2b08fd0d9c44" (UID: "cae5c530-0398-48a0-ad06-2b08fd0d9c44"). InnerVolumeSpecName "kube-api-access-2nvzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.501829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-config-data" (OuterVolumeSpecName: "config-data") pod "cae5c530-0398-48a0-ad06-2b08fd0d9c44" (UID: "cae5c530-0398-48a0-ad06-2b08fd0d9c44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.508847 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cae5c530-0398-48a0-ad06-2b08fd0d9c44" (UID: "cae5c530-0398-48a0-ad06-2b08fd0d9c44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.572150 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nvzx\" (UniqueName: \"kubernetes.io/projected/cae5c530-0398-48a0-ad06-2b08fd0d9c44-kube-api-access-2nvzx\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.572198 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.572212 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:13 crc kubenswrapper[4809]: I1205 11:33:13.572225 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5c530-0398-48a0-ad06-2b08fd0d9c44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.046888 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.046943 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.064597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-qnwjc" event={"ID":"cae5c530-0398-48a0-ad06-2b08fd0d9c44","Type":"ContainerDied","Data":"d085dd685d14158532bd2442d264320133a43b843ed67c8bb19c4314bbd03459"} Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.064668 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d085dd685d14158532bd2442d264320133a43b843ed67c8bb19c4314bbd03459" Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.064608 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-qnwjc" Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.290587 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.291147 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-log" containerID="cri-o://fc8f6fbd0d6af548b3ab28ac3220e10e5ec059642c2119b7cb235f5f82975b77" gracePeriod=30 Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.291497 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-api" containerID="cri-o://99c7a952c43dee727c2d091290d87f6671731e8e5972b24248e47525e3da6a3a" gracePeriod=30 Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.307011 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.307218 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="18669644-e182-493a-8a14-2625f80e7e1f" containerName="nova-scheduler-scheduler" containerID="cri-o://65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d" gracePeriod=30 Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.308940 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.309036 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.354550 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.354799 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-log" containerID="cri-o://187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa" gracePeriod=30 Dec 05 11:33:14 crc kubenswrapper[4809]: I1205 11:33:14.354883 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-metadata" containerID="cri-o://2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e" gracePeriod=30 Dec 05 11:33:15 crc kubenswrapper[4809]: I1205 11:33:15.074540 4809 generic.go:334] "Generic (PLEG): container finished" podID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerID="187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa" exitCode=143 Dec 05 11:33:15 crc kubenswrapper[4809]: I1205 11:33:15.074646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c069cd4-4bde-4705-afa0-8330c0edd383","Type":"ContainerDied","Data":"187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa"} Dec 05 11:33:15 crc kubenswrapper[4809]: I1205 11:33:15.076589 4809 generic.go:334] "Generic (PLEG): container finished" podID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerID="fc8f6fbd0d6af548b3ab28ac3220e10e5ec059642c2119b7cb235f5f82975b77" exitCode=143 Dec 05 11:33:15 crc kubenswrapper[4809]: I1205 11:33:15.076622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83a4702d-662e-4c2b-8335-d5f43099c44b","Type":"ContainerDied","Data":"fc8f6fbd0d6af548b3ab28ac3220e10e5ec059642c2119b7cb235f5f82975b77"} Dec 05 11:33:15 crc kubenswrapper[4809]: E1205 11:33:15.277956 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:33:15 crc kubenswrapper[4809]: E1205 11:33:15.290003 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:33:15 crc kubenswrapper[4809]: E1205 11:33:15.294915 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:33:15 crc kubenswrapper[4809]: E1205 11:33:15.295367 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="18669644-e182-493a-8a14-2625f80e7e1f" containerName="nova-scheduler-scheduler" Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.536612 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:53790->10.217.0.192:8775: read: connection reset by peer" Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.536715 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:53774->10.217.0.192:8775: read: connection reset by peer" Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.943911 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.962743 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-combined-ca-bundle\") pod \"4c069cd4-4bde-4705-afa0-8330c0edd383\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.962864 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-config-data\") pod \"4c069cd4-4bde-4705-afa0-8330c0edd383\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.962930 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c069cd4-4bde-4705-afa0-8330c0edd383-logs\") pod \"4c069cd4-4bde-4705-afa0-8330c0edd383\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.962990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-nova-metadata-tls-certs\") pod \"4c069cd4-4bde-4705-afa0-8330c0edd383\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.963019 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dthrw\" (UniqueName: \"kubernetes.io/projected/4c069cd4-4bde-4705-afa0-8330c0edd383-kube-api-access-dthrw\") pod \"4c069cd4-4bde-4705-afa0-8330c0edd383\" (UID: \"4c069cd4-4bde-4705-afa0-8330c0edd383\") " Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.963549 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c069cd4-4bde-4705-afa0-8330c0edd383-logs" (OuterVolumeSpecName: "logs") pod "4c069cd4-4bde-4705-afa0-8330c0edd383" (UID: "4c069cd4-4bde-4705-afa0-8330c0edd383"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.963843 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c069cd4-4bde-4705-afa0-8330c0edd383-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:17 crc kubenswrapper[4809]: I1205 11:33:17.973942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c069cd4-4bde-4705-afa0-8330c0edd383-kube-api-access-dthrw" (OuterVolumeSpecName: "kube-api-access-dthrw") pod "4c069cd4-4bde-4705-afa0-8330c0edd383" (UID: "4c069cd4-4bde-4705-afa0-8330c0edd383"). InnerVolumeSpecName "kube-api-access-dthrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.017106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-config-data" (OuterVolumeSpecName: "config-data") pod "4c069cd4-4bde-4705-afa0-8330c0edd383" (UID: "4c069cd4-4bde-4705-afa0-8330c0edd383"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.034341 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c069cd4-4bde-4705-afa0-8330c0edd383" (UID: "4c069cd4-4bde-4705-afa0-8330c0edd383"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.066268 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.066307 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dthrw\" (UniqueName: \"kubernetes.io/projected/4c069cd4-4bde-4705-afa0-8330c0edd383-kube-api-access-dthrw\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.066320 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.068910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4c069cd4-4bde-4705-afa0-8330c0edd383" (UID: "4c069cd4-4bde-4705-afa0-8330c0edd383"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.116866 4809 generic.go:334] "Generic (PLEG): container finished" podID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerID="2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e" exitCode=0 Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.116917 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c069cd4-4bde-4705-afa0-8330c0edd383","Type":"ContainerDied","Data":"2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e"} Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.116955 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c069cd4-4bde-4705-afa0-8330c0edd383","Type":"ContainerDied","Data":"bc1ee4cfa87b4b8bbaf243ee9c3f72154fcd54d2fc9b5b547752ece32d61fe83"} Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.117013 4809 scope.go:117] "RemoveContainer" containerID="2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.117228 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.138282 4809 scope.go:117] "RemoveContainer" containerID="187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.157429 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.169723 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.175619 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:33:18 crc kubenswrapper[4809]: E1205 11:33:18.176051 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-log" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.176068 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-log" Dec 05 11:33:18 crc kubenswrapper[4809]: E1205 11:33:18.176092 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae5c530-0398-48a0-ad06-2b08fd0d9c44" containerName="nova-manage" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.176099 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae5c530-0398-48a0-ad06-2b08fd0d9c44" containerName="nova-manage" Dec 05 11:33:18 crc kubenswrapper[4809]: E1205 11:33:18.176115 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-metadata" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.176121 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-metadata" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.176277 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae5c530-0398-48a0-ad06-2b08fd0d9c44" containerName="nova-manage" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.176292 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-log" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.176298 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" containerName="nova-metadata-metadata" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.177325 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.178047 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c069cd4-4bde-4705-afa0-8330c0edd383-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.179888 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.179979 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.199569 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.207775 4809 scope.go:117] "RemoveContainer" containerID="2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e" Dec 05 11:33:18 crc kubenswrapper[4809]: E1205 11:33:18.217648 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e\": container with ID starting with 2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e not found: ID does not exist" containerID="2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.217693 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e"} err="failed to get container status \"2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e\": rpc error: code = NotFound desc = could not find container \"2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e\": container with ID starting with 2980fff744017b8c26fa31fa3cdabb5927fa1702a19c76ef6ef08adde8f4118e not found: ID does not exist" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.217735 4809 scope.go:117] "RemoveContainer" containerID="187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa" Dec 05 11:33:18 crc kubenswrapper[4809]: E1205 11:33:18.218883 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa\": container with ID starting with 187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa not found: ID does not exist" containerID="187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.219077 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa"} err="failed to get container status \"187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa\": rpc error: code = NotFound desc = could not find container \"187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa\": container with ID starting with 187e2d1e98a018d7a78e2b45bc5154319b6d99849aee4da416c5ac4322a5b9fa not found: ID does not exist" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.279169 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.279450 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871c924c-f749-40cd-81b2-64af559e6b20-logs\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.279750 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mnlp\" (UniqueName: \"kubernetes.io/projected/871c924c-f749-40cd-81b2-64af559e6b20-kube-api-access-5mnlp\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.279952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-config-data\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.280202 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.382021 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.382072 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871c924c-f749-40cd-81b2-64af559e6b20-logs\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.382145 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mnlp\" (UniqueName: \"kubernetes.io/projected/871c924c-f749-40cd-81b2-64af559e6b20-kube-api-access-5mnlp\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.382181 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-config-data\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.382239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.383024 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871c924c-f749-40cd-81b2-64af559e6b20-logs\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.386065 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.387623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-config-data\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.401121 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.433224 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mnlp\" (UniqueName: \"kubernetes.io/projected/871c924c-f749-40cd-81b2-64af559e6b20-kube-api-access-5mnlp\") pod \"nova-metadata-0\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.508843 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:33:18 crc kubenswrapper[4809]: I1205 11:33:18.892699 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c069cd4-4bde-4705-afa0-8330c0edd383" path="/var/lib/kubelet/pods/4c069cd4-4bde-4705-afa0-8330c0edd383/volumes" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.000108 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:33:19 crc kubenswrapper[4809]: W1205 11:33:19.031773 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod871c924c_f749_40cd_81b2_64af559e6b20.slice/crio-ecffc19de760f93ad003eed298aef231e8504b94d7049fff23cc61e0f2b3f211 WatchSource:0}: Error finding container ecffc19de760f93ad003eed298aef231e8504b94d7049fff23cc61e0f2b3f211: Status 404 returned error can't find the container with id ecffc19de760f93ad003eed298aef231e8504b94d7049fff23cc61e0f2b3f211 Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.128991 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"871c924c-f749-40cd-81b2-64af559e6b20","Type":"ContainerStarted","Data":"ecffc19de760f93ad003eed298aef231e8504b94d7049fff23cc61e0f2b3f211"} Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.131137 4809 generic.go:334] "Generic (PLEG): container finished" podID="18669644-e182-493a-8a14-2625f80e7e1f" containerID="65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d" exitCode=0 Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.131215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18669644-e182-493a-8a14-2625f80e7e1f","Type":"ContainerDied","Data":"65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d"} Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.176011 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bzgb8"] Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.179309 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.219599 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bzgb8"] Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.314999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-utilities\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.315408 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-catalog-content\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.315652 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8swd\" (UniqueName: \"kubernetes.io/projected/621517a1-ab52-4471-979a-bee6f0ab8311-kube-api-access-p8swd\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.333150 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.419969 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdgfl\" (UniqueName: \"kubernetes.io/projected/18669644-e182-493a-8a14-2625f80e7e1f-kube-api-access-kdgfl\") pod \"18669644-e182-493a-8a14-2625f80e7e1f\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.420089 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-config-data\") pod \"18669644-e182-493a-8a14-2625f80e7e1f\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.420280 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-combined-ca-bundle\") pod \"18669644-e182-493a-8a14-2625f80e7e1f\" (UID: \"18669644-e182-493a-8a14-2625f80e7e1f\") " Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.420889 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-catalog-content\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.421037 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8swd\" (UniqueName: \"kubernetes.io/projected/621517a1-ab52-4471-979a-bee6f0ab8311-kube-api-access-p8swd\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.421136 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-utilities\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.421533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-catalog-content\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.421602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-utilities\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.423991 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18669644-e182-493a-8a14-2625f80e7e1f-kube-api-access-kdgfl" (OuterVolumeSpecName: "kube-api-access-kdgfl") pod "18669644-e182-493a-8a14-2625f80e7e1f" (UID: "18669644-e182-493a-8a14-2625f80e7e1f"). InnerVolumeSpecName "kube-api-access-kdgfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.441507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8swd\" (UniqueName: \"kubernetes.io/projected/621517a1-ab52-4471-979a-bee6f0ab8311-kube-api-access-p8swd\") pod \"redhat-operators-bzgb8\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.455795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-config-data" (OuterVolumeSpecName: "config-data") pod "18669644-e182-493a-8a14-2625f80e7e1f" (UID: "18669644-e182-493a-8a14-2625f80e7e1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.458574 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18669644-e182-493a-8a14-2625f80e7e1f" (UID: "18669644-e182-493a-8a14-2625f80e7e1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.520553 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.522536 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.522650 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18669644-e182-493a-8a14-2625f80e7e1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.522715 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdgfl\" (UniqueName: \"kubernetes.io/projected/18669644-e182-493a-8a14-2625f80e7e1f-kube-api-access-kdgfl\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:19 crc kubenswrapper[4809]: I1205 11:33:19.973272 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bzgb8"] Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.144702 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"871c924c-f749-40cd-81b2-64af559e6b20","Type":"ContainerStarted","Data":"6e8631852c5ee2d2ff12bc4e2bff28c710f77be32894780bc515ed6be7263ddd"} Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.144962 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"871c924c-f749-40cd-81b2-64af559e6b20","Type":"ContainerStarted","Data":"65c7a2783b79fc201d707bcdea31bba4d9626dc6a1edc4ed23ea125b15d958bf"} Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.147675 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzgb8" event={"ID":"621517a1-ab52-4471-979a-bee6f0ab8311","Type":"ContainerStarted","Data":"b5883d02b8d991b0276830e08da342820e08e4982b05740a55967901bf8cfa48"} Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.150141 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18669644-e182-493a-8a14-2625f80e7e1f","Type":"ContainerDied","Data":"eb93806dc4c131e3bc3a4e5eb5eba24f58e821d3a036dbae82fd76aaea8f5a15"} Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.150175 4809 scope.go:117] "RemoveContainer" containerID="65cd273a101140eae0f480d2aeade8c332f32ec0273a194879f400f96c6df80d" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.150176 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.153145 4809 generic.go:334] "Generic (PLEG): container finished" podID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerID="99c7a952c43dee727c2d091290d87f6671731e8e5972b24248e47525e3da6a3a" exitCode=0 Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.153170 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83a4702d-662e-4c2b-8335-d5f43099c44b","Type":"ContainerDied","Data":"99c7a952c43dee727c2d091290d87f6671731e8e5972b24248e47525e3da6a3a"} Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.153207 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83a4702d-662e-4c2b-8335-d5f43099c44b","Type":"ContainerDied","Data":"67b3fe0142c80e79a10796ea7321b89266b3f8a6bdfbd2e8e8ec2e4edea4fc6a"} Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.153220 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67b3fe0142c80e79a10796ea7321b89266b3f8a6bdfbd2e8e8ec2e4edea4fc6a" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.169346 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.169331066 podStartE2EDuration="2.169331066s" podCreationTimestamp="2025-12-05 11:33:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:33:20.166655277 +0000 UTC m=+1495.557631855" watchObservedRunningTime="2025-12-05 11:33:20.169331066 +0000 UTC m=+1495.560307624" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.182414 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.205317 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.220766 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.234676 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83a4702d-662e-4c2b-8335-d5f43099c44b-logs\") pod \"83a4702d-662e-4c2b-8335-d5f43099c44b\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.234761 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-combined-ca-bundle\") pod \"83a4702d-662e-4c2b-8335-d5f43099c44b\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.234827 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-public-tls-certs\") pod \"83a4702d-662e-4c2b-8335-d5f43099c44b\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.234893 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5snhb\" (UniqueName: \"kubernetes.io/projected/83a4702d-662e-4c2b-8335-d5f43099c44b-kube-api-access-5snhb\") pod \"83a4702d-662e-4c2b-8335-d5f43099c44b\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.234951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-config-data\") pod \"83a4702d-662e-4c2b-8335-d5f43099c44b\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.234992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-internal-tls-certs\") pod \"83a4702d-662e-4c2b-8335-d5f43099c44b\" (UID: \"83a4702d-662e-4c2b-8335-d5f43099c44b\") " Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.235106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a4702d-662e-4c2b-8335-d5f43099c44b-logs" (OuterVolumeSpecName: "logs") pod "83a4702d-662e-4c2b-8335-d5f43099c44b" (UID: "83a4702d-662e-4c2b-8335-d5f43099c44b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.235554 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83a4702d-662e-4c2b-8335-d5f43099c44b-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.238219 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:33:20 crc kubenswrapper[4809]: E1205 11:33:20.238642 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-api" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.238667 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-api" Dec 05 11:33:20 crc kubenswrapper[4809]: E1205 11:33:20.238701 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18669644-e182-493a-8a14-2625f80e7e1f" containerName="nova-scheduler-scheduler" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.238707 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="18669644-e182-493a-8a14-2625f80e7e1f" containerName="nova-scheduler-scheduler" Dec 05 11:33:20 crc kubenswrapper[4809]: E1205 11:33:20.238716 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-log" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.238722 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-log" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.238918 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="18669644-e182-493a-8a14-2625f80e7e1f" containerName="nova-scheduler-scheduler" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.238934 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-api" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.238949 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" containerName="nova-api-log" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.239703 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.243105 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a4702d-662e-4c2b-8335-d5f43099c44b-kube-api-access-5snhb" (OuterVolumeSpecName: "kube-api-access-5snhb") pod "83a4702d-662e-4c2b-8335-d5f43099c44b" (UID: "83a4702d-662e-4c2b-8335-d5f43099c44b"). InnerVolumeSpecName "kube-api-access-5snhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.259019 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.261836 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.280297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83a4702d-662e-4c2b-8335-d5f43099c44b" (UID: "83a4702d-662e-4c2b-8335-d5f43099c44b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.308747 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-config-data" (OuterVolumeSpecName: "config-data") pod "83a4702d-662e-4c2b-8335-d5f43099c44b" (UID: "83a4702d-662e-4c2b-8335-d5f43099c44b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.337800 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-config-data\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.337843 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwlkw\" (UniqueName: \"kubernetes.io/projected/cc828bad-c058-4a1c-aa44-18d607759d6b-kube-api-access-dwlkw\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.338129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.338465 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.338487 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.338501 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5snhb\" (UniqueName: \"kubernetes.io/projected/83a4702d-662e-4c2b-8335-d5f43099c44b-kube-api-access-5snhb\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.346975 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "83a4702d-662e-4c2b-8335-d5f43099c44b" (UID: "83a4702d-662e-4c2b-8335-d5f43099c44b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.359613 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "83a4702d-662e-4c2b-8335-d5f43099c44b" (UID: "83a4702d-662e-4c2b-8335-d5f43099c44b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.440315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.440439 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-config-data\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.440460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwlkw\" (UniqueName: \"kubernetes.io/projected/cc828bad-c058-4a1c-aa44-18d607759d6b-kube-api-access-dwlkw\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.440538 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.440548 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a4702d-662e-4c2b-8335-d5f43099c44b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.443864 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-config-data\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.445125 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.459306 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwlkw\" (UniqueName: \"kubernetes.io/projected/cc828bad-c058-4a1c-aa44-18d607759d6b-kube-api-access-dwlkw\") pod \"nova-scheduler-0\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.577074 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:33:20 crc kubenswrapper[4809]: I1205 11:33:20.886001 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18669644-e182-493a-8a14-2625f80e7e1f" path="/var/lib/kubelet/pods/18669644-e182-493a-8a14-2625f80e7e1f/volumes" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.053272 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:33:21 crc kubenswrapper[4809]: W1205 11:33:21.057892 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc828bad_c058_4a1c_aa44_18d607759d6b.slice/crio-37dc18c66a331db4a85e69a97c224a913b4d46939aab7b1a450f853642d114fe WatchSource:0}: Error finding container 37dc18c66a331db4a85e69a97c224a913b4d46939aab7b1a450f853642d114fe: Status 404 returned error can't find the container with id 37dc18c66a331db4a85e69a97c224a913b4d46939aab7b1a450f853642d114fe Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.168825 4809 generic.go:334] "Generic (PLEG): container finished" podID="621517a1-ab52-4471-979a-bee6f0ab8311" containerID="fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942" exitCode=0 Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.168914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzgb8" event={"ID":"621517a1-ab52-4471-979a-bee6f0ab8311","Type":"ContainerDied","Data":"fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942"} Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.172218 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc828bad-c058-4a1c-aa44-18d607759d6b","Type":"ContainerStarted","Data":"37dc18c66a331db4a85e69a97c224a913b4d46939aab7b1a450f853642d114fe"} Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.184102 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.253815 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.264013 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.273371 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.276304 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.279404 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.279668 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.279762 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.285413 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.357366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-config-data\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.357491 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-logs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.357579 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.357882 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khlmm\" (UniqueName: \"kubernetes.io/projected/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-kube-api-access-khlmm\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.357989 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-public-tls-certs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.358046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.460466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-logs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.460663 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.460759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khlmm\" (UniqueName: \"kubernetes.io/projected/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-kube-api-access-khlmm\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.460816 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-public-tls-certs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.460891 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.460969 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-config-data\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.461677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-logs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.466296 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-config-data\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.467332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.468421 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-public-tls-certs\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.472895 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.477325 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khlmm\" (UniqueName: \"kubernetes.io/projected/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-kube-api-access-khlmm\") pod \"nova-api-0\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.600341 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:33:21 crc kubenswrapper[4809]: I1205 11:33:21.899432 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:33:22 crc kubenswrapper[4809]: I1205 11:33:22.225672 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc828bad-c058-4a1c-aa44-18d607759d6b","Type":"ContainerStarted","Data":"1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2"} Dec 05 11:33:22 crc kubenswrapper[4809]: I1205 11:33:22.248792 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8220cf04-d91a-45d8-ad53-51d0c6d21fe5","Type":"ContainerStarted","Data":"063a6dfa4af4111732f2aa27c4f351895860949b0c9d245dfe75cbda20b6e1cc"} Dec 05 11:33:22 crc kubenswrapper[4809]: I1205 11:33:22.248846 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8220cf04-d91a-45d8-ad53-51d0c6d21fe5","Type":"ContainerStarted","Data":"a22bd508ef19c64780c063829731f0660026372c8ade33583c9a9ecd7c2a2435"} Dec 05 11:33:22 crc kubenswrapper[4809]: I1205 11:33:22.268852 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.268832193 podStartE2EDuration="2.268832193s" podCreationTimestamp="2025-12-05 11:33:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:33:22.254961195 +0000 UTC m=+1497.645937763" watchObservedRunningTime="2025-12-05 11:33:22.268832193 +0000 UTC m=+1497.659808761" Dec 05 11:33:22 crc kubenswrapper[4809]: I1205 11:33:22.903594 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a4702d-662e-4c2b-8335-d5f43099c44b" path="/var/lib/kubelet/pods/83a4702d-662e-4c2b-8335-d5f43099c44b/volumes" Dec 05 11:33:23 crc kubenswrapper[4809]: I1205 11:33:23.263587 4809 generic.go:334] "Generic (PLEG): container finished" podID="621517a1-ab52-4471-979a-bee6f0ab8311" containerID="3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b" exitCode=0 Dec 05 11:33:23 crc kubenswrapper[4809]: I1205 11:33:23.263841 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzgb8" event={"ID":"621517a1-ab52-4471-979a-bee6f0ab8311","Type":"ContainerDied","Data":"3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b"} Dec 05 11:33:23 crc kubenswrapper[4809]: I1205 11:33:23.269223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8220cf04-d91a-45d8-ad53-51d0c6d21fe5","Type":"ContainerStarted","Data":"2e48dcfd66c9332d451a5a78ac5b9431515f09a61cddb083325ef0de295b6608"} Dec 05 11:33:23 crc kubenswrapper[4809]: I1205 11:33:23.327203 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.327176528 podStartE2EDuration="2.327176528s" podCreationTimestamp="2025-12-05 11:33:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:33:23.30590296 +0000 UTC m=+1498.696879518" watchObservedRunningTime="2025-12-05 11:33:23.327176528 +0000 UTC m=+1498.718153126" Dec 05 11:33:23 crc kubenswrapper[4809]: I1205 11:33:23.509210 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:33:23 crc kubenswrapper[4809]: I1205 11:33:23.509490 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:33:25 crc kubenswrapper[4809]: I1205 11:33:25.290139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzgb8" event={"ID":"621517a1-ab52-4471-979a-bee6f0ab8311","Type":"ContainerStarted","Data":"a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89"} Dec 05 11:33:25 crc kubenswrapper[4809]: I1205 11:33:25.329454 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bzgb8" podStartSLOduration=2.554665457 podStartE2EDuration="6.329434277s" podCreationTimestamp="2025-12-05 11:33:19 +0000 UTC" firstStartedPulling="2025-12-05 11:33:21.171312676 +0000 UTC m=+1496.562289234" lastFinishedPulling="2025-12-05 11:33:24.946081496 +0000 UTC m=+1500.337058054" observedRunningTime="2025-12-05 11:33:25.314101241 +0000 UTC m=+1500.705077819" watchObservedRunningTime="2025-12-05 11:33:25.329434277 +0000 UTC m=+1500.720410835" Dec 05 11:33:25 crc kubenswrapper[4809]: I1205 11:33:25.578217 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 11:33:28 crc kubenswrapper[4809]: I1205 11:33:28.510055 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:33:28 crc kubenswrapper[4809]: I1205 11:33:28.510419 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:33:29 crc kubenswrapper[4809]: I1205 11:33:29.521624 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:29 crc kubenswrapper[4809]: I1205 11:33:29.522124 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:29 crc kubenswrapper[4809]: I1205 11:33:29.524932 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:33:29 crc kubenswrapper[4809]: I1205 11:33:29.525674 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:33:30 crc kubenswrapper[4809]: I1205 11:33:30.566268 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bzgb8" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="registry-server" probeResult="failure" output=< Dec 05 11:33:30 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 11:33:30 crc kubenswrapper[4809]: > Dec 05 11:33:30 crc kubenswrapper[4809]: I1205 11:33:30.578126 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 11:33:30 crc kubenswrapper[4809]: I1205 11:33:30.759473 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 11:33:31 crc kubenswrapper[4809]: I1205 11:33:31.408326 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 11:33:31 crc kubenswrapper[4809]: I1205 11:33:31.601765 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:33:31 crc kubenswrapper[4809]: I1205 11:33:31.601827 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:33:32 crc kubenswrapper[4809]: I1205 11:33:32.616962 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:33:32 crc kubenswrapper[4809]: I1205 11:33:32.617006 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:33:38 crc kubenswrapper[4809]: I1205 11:33:38.427270 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 11:33:38 crc kubenswrapper[4809]: I1205 11:33:38.536692 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 11:33:38 crc kubenswrapper[4809]: I1205 11:33:38.536762 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 11:33:38 crc kubenswrapper[4809]: I1205 11:33:38.546717 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 11:33:38 crc kubenswrapper[4809]: I1205 11:33:38.556462 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 11:33:39 crc kubenswrapper[4809]: I1205 11:33:39.589021 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:39 crc kubenswrapper[4809]: I1205 11:33:39.642500 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:39 crc kubenswrapper[4809]: I1205 11:33:39.837190 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bzgb8"] Dec 05 11:33:41 crc kubenswrapper[4809]: I1205 11:33:41.449654 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bzgb8" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="registry-server" containerID="cri-o://a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89" gracePeriod=2 Dec 05 11:33:41 crc kubenswrapper[4809]: I1205 11:33:41.617487 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:33:41 crc kubenswrapper[4809]: I1205 11:33:41.618434 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:33:41 crc kubenswrapper[4809]: I1205 11:33:41.619274 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:33:41 crc kubenswrapper[4809]: I1205 11:33:41.639347 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:33:41 crc kubenswrapper[4809]: I1205 11:33:41.949481 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.089454 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-catalog-content\") pod \"621517a1-ab52-4471-979a-bee6f0ab8311\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.089699 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8swd\" (UniqueName: \"kubernetes.io/projected/621517a1-ab52-4471-979a-bee6f0ab8311-kube-api-access-p8swd\") pod \"621517a1-ab52-4471-979a-bee6f0ab8311\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.089979 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-utilities\") pod \"621517a1-ab52-4471-979a-bee6f0ab8311\" (UID: \"621517a1-ab52-4471-979a-bee6f0ab8311\") " Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.090553 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-utilities" (OuterVolumeSpecName: "utilities") pod "621517a1-ab52-4471-979a-bee6f0ab8311" (UID: "621517a1-ab52-4471-979a-bee6f0ab8311"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.095725 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/621517a1-ab52-4471-979a-bee6f0ab8311-kube-api-access-p8swd" (OuterVolumeSpecName: "kube-api-access-p8swd") pod "621517a1-ab52-4471-979a-bee6f0ab8311" (UID: "621517a1-ab52-4471-979a-bee6f0ab8311"). InnerVolumeSpecName "kube-api-access-p8swd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.192580 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.192603 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8swd\" (UniqueName: \"kubernetes.io/projected/621517a1-ab52-4471-979a-bee6f0ab8311-kube-api-access-p8swd\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.195277 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "621517a1-ab52-4471-979a-bee6f0ab8311" (UID: "621517a1-ab52-4471-979a-bee6f0ab8311"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.294180 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/621517a1-ab52-4471-979a-bee6f0ab8311-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.468697 4809 generic.go:334] "Generic (PLEG): container finished" podID="621517a1-ab52-4471-979a-bee6f0ab8311" containerID="a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89" exitCode=0 Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.468802 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzgb8" event={"ID":"621517a1-ab52-4471-979a-bee6f0ab8311","Type":"ContainerDied","Data":"a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89"} Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.468856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bzgb8" event={"ID":"621517a1-ab52-4471-979a-bee6f0ab8311","Type":"ContainerDied","Data":"b5883d02b8d991b0276830e08da342820e08e4982b05740a55967901bf8cfa48"} Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.468855 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bzgb8" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.468883 4809 scope.go:117] "RemoveContainer" containerID="a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.469359 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.476845 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.529620 4809 scope.go:117] "RemoveContainer" containerID="3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.542687 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bzgb8"] Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.556555 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bzgb8"] Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.580878 4809 scope.go:117] "RemoveContainer" containerID="fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.621858 4809 scope.go:117] "RemoveContainer" containerID="a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89" Dec 05 11:33:42 crc kubenswrapper[4809]: E1205 11:33:42.622731 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89\": container with ID starting with a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89 not found: ID does not exist" containerID="a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.622756 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89"} err="failed to get container status \"a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89\": rpc error: code = NotFound desc = could not find container \"a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89\": container with ID starting with a55cce6686a0e5f878b49efdc065f2bf89692d5534cdc8bb8b744e7719fd0d89 not found: ID does not exist" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.622777 4809 scope.go:117] "RemoveContainer" containerID="3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b" Dec 05 11:33:42 crc kubenswrapper[4809]: E1205 11:33:42.623064 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b\": container with ID starting with 3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b not found: ID does not exist" containerID="3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.623079 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b"} err="failed to get container status \"3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b\": rpc error: code = NotFound desc = could not find container \"3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b\": container with ID starting with 3fde6729602e3c5a7a3b8f6276be954c813f5f2c4d30fbb1d9701e7c9648fb9b not found: ID does not exist" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.623092 4809 scope.go:117] "RemoveContainer" containerID="fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942" Dec 05 11:33:42 crc kubenswrapper[4809]: E1205 11:33:42.623359 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942\": container with ID starting with fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942 not found: ID does not exist" containerID="fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.623374 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942"} err="failed to get container status \"fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942\": rpc error: code = NotFound desc = could not find container \"fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942\": container with ID starting with fcf9543e57ebb1d7266502961322e27d39ab86488262070d617df300e1d4a942 not found: ID does not exist" Dec 05 11:33:42 crc kubenswrapper[4809]: I1205 11:33:42.889255 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" path="/var/lib/kubelet/pods/621517a1-ab52-4471-979a-bee6f0ab8311/volumes" Dec 05 11:33:44 crc kubenswrapper[4809]: I1205 11:33:44.046445 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:33:44 crc kubenswrapper[4809]: I1205 11:33:44.046863 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:34:02 crc kubenswrapper[4809]: I1205 11:34:02.914601 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 11:34:02 crc kubenswrapper[4809]: I1205 11:34:02.915363 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" containerName="openstackclient" containerID="cri-o://719fed1b9e606f1acf047af8258117088542a8ab9a3b33d882683eb029e590fd" gracePeriod=2 Dec 05 11:34:02 crc kubenswrapper[4809]: I1205 11:34:02.931850 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.222829 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementef3a-account-delete-qhc9d"] Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.225606 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="extract-utilities" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.225828 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="extract-utilities" Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.225919 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="registry-server" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.226000 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="registry-server" Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.226131 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" containerName="openstackclient" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.226213 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" containerName="openstackclient" Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.226321 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="extract-content" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.226397 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="extract-content" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.226815 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="621517a1-ab52-4471-979a-bee6f0ab8311" containerName="registry-server" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.226935 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" containerName="openstackclient" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.227793 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.241972 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementef3a-account-delete-qhc9d"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.307003 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance21cd-account-delete-kthxx"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.308144 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.333334 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance21cd-account-delete-kthxx"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.356453 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.376693 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/447a5847-564c-4b01-a74d-503249e0dd95-operator-scripts\") pod \"placementef3a-account-delete-qhc9d\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.376753 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2f29\" (UniqueName: \"kubernetes.io/projected/447a5847-564c-4b01-a74d-503249e0dd95-kube-api-access-n2f29\") pod \"placementef3a-account-delete-qhc9d\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.376814 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6712dcd8-a4a4-4f57-adee-6b218823ba77-operator-scripts\") pod \"glance21cd-account-delete-kthxx\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.376870 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qftns\" (UniqueName: \"kubernetes.io/projected/6712dcd8-a4a4-4f57-adee-6b218823ba77-kube-api-access-qftns\") pod \"glance21cd-account-delete-kthxx\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.382155 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.382723 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="openstack-network-exporter" containerID="cri-o://08a2f85174278c931e6dfac745c0fe3ac459b86ac3c72c047078366ce8f1b112" gracePeriod=300 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.477674 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="ovsdbserver-sb" containerID="cri-o://f754ccd26270d79dad0c1628da75aaa14922d40a673887d3fffac5fd152c8685" gracePeriod=300 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.478575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6712dcd8-a4a4-4f57-adee-6b218823ba77-operator-scripts\") pod \"glance21cd-account-delete-kthxx\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.478698 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qftns\" (UniqueName: \"kubernetes.io/projected/6712dcd8-a4a4-4f57-adee-6b218823ba77-kube-api-access-qftns\") pod \"glance21cd-account-delete-kthxx\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.478769 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/447a5847-564c-4b01-a74d-503249e0dd95-operator-scripts\") pod \"placementef3a-account-delete-qhc9d\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.478822 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2f29\" (UniqueName: \"kubernetes.io/projected/447a5847-564c-4b01-a74d-503249e0dd95-kube-api-access-n2f29\") pod \"placementef3a-account-delete-qhc9d\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.480619 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6712dcd8-a4a4-4f57-adee-6b218823ba77-operator-scripts\") pod \"glance21cd-account-delete-kthxx\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.480676 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/447a5847-564c-4b01-a74d-503249e0dd95-operator-scripts\") pod \"placementef3a-account-delete-qhc9d\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.480876 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.481007 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data podName:e1251531-cb99-418e-aa38-65a793e5a9d0 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:03.980991164 +0000 UTC m=+1539.371967742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data") pod "rabbitmq-server-0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0") : configmap "rabbitmq-config-data" not found Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.516310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2f29\" (UniqueName: \"kubernetes.io/projected/447a5847-564c-4b01-a74d-503249e0dd95-kube-api-access-n2f29\") pod \"placementef3a-account-delete-qhc9d\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.516384 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qftns\" (UniqueName: \"kubernetes.io/projected/6712dcd8-a4a4-4f57-adee-6b218823ba77-kube-api-access-qftns\") pod \"glance21cd-account-delete-kthxx\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.562842 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder18b7-account-delete-jcsnx"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.564084 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.565454 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.577524 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder18b7-account-delete-jcsnx"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.636475 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.670478 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronee61-account-delete-dfhwt"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.671622 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.686926 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk9s6\" (UniqueName: \"kubernetes.io/projected/102c9fa9-707d-4918-9a5d-55d8d4c47c12-kube-api-access-zk9s6\") pod \"cinder18b7-account-delete-jcsnx\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.686965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts\") pod \"cinder18b7-account-delete-jcsnx\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.693882 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronee61-account-delete-dfhwt"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.741642 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.742001 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="ovn-northd" containerID="cri-o://043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" gracePeriod=30 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.742116 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="openstack-network-exporter" containerID="cri-o://855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8" gracePeriod=30 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.758208 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.759109 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="openstack-network-exporter" containerID="cri-o://17bf3535a2282730bf4584649cdef0fdc6c738363db31b9323fc6bba727833db" gracePeriod=300 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.789359 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk9s6\" (UniqueName: \"kubernetes.io/projected/102c9fa9-707d-4918-9a5d-55d8d4c47c12-kube-api-access-zk9s6\") pod \"cinder18b7-account-delete-jcsnx\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.789403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts\") pod \"cinder18b7-account-delete-jcsnx\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.789458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hn6s\" (UniqueName: \"kubernetes.io/projected/77f68706-76e6-4ee3-b27d-bb092ab68a30-kube-api-access-5hn6s\") pod \"neutronee61-account-delete-dfhwt\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.789507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts\") pod \"neutronee61-account-delete-dfhwt\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.796338 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts\") pod \"cinder18b7-account-delete-jcsnx\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.814512 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ddjc8"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.827171 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican92b5-account-delete-x25h6"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.828668 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.835805 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ddjc8"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.845680 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican92b5-account-delete-x25h6"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.851035 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fd04fa46-c62b-4b8e-947d-da5846355feb/ovsdbserver-sb/0.log" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.851081 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerID="08a2f85174278c931e6dfac745c0fe3ac459b86ac3c72c047078366ce8f1b112" exitCode=2 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.851095 4809 generic.go:334] "Generic (PLEG): container finished" podID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerID="f754ccd26270d79dad0c1628da75aaa14922d40a673887d3fffac5fd152c8685" exitCode=143 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.851114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fd04fa46-c62b-4b8e-947d-da5846355feb","Type":"ContainerDied","Data":"08a2f85174278c931e6dfac745c0fe3ac459b86ac3c72c047078366ce8f1b112"} Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.851137 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fd04fa46-c62b-4b8e-947d-da5846355feb","Type":"ContainerDied","Data":"f754ccd26270d79dad0c1628da75aaa14922d40a673887d3fffac5fd152c8685"} Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.876058 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder-api-0" secret="" err="secret \"cinder-cinder-dockercfg-q2l2q\" not found" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.879286 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-5mwp9"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.889836 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk9s6\" (UniqueName: \"kubernetes.io/projected/102c9fa9-707d-4918-9a5d-55d8d4c47c12-kube-api-access-zk9s6\") pod \"cinder18b7-account-delete-jcsnx\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.890859 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts\") pod \"neutronee61-account-delete-dfhwt\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.890944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-operator-scripts\") pod \"barbican92b5-account-delete-x25h6\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.891043 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbb25\" (UniqueName: \"kubernetes.io/projected/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-kube-api-access-vbb25\") pod \"barbican92b5-account-delete-x25h6\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.891088 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hn6s\" (UniqueName: \"kubernetes.io/projected/77f68706-76e6-4ee3-b27d-bb092ab68a30-kube-api-access-5hn6s\") pod \"neutronee61-account-delete-dfhwt\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.893571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts\") pod \"neutronee61-account-delete-dfhwt\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.894042 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-5mwp9"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.899058 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="ovsdbserver-nb" containerID="cri-o://0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a" gracePeriod=300 Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.902709 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-knlr7"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.916820 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-knlr7"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.923369 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hn6s\" (UniqueName: \"kubernetes.io/projected/77f68706-76e6-4ee3-b27d-bb092ab68a30-kube-api-access-5hn6s\") pod \"neutronee61-account-delete-dfhwt\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.956818 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-9mn2t"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.983608 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-9mn2t"] Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.995754 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-operator-scripts\") pod \"barbican92b5-account-delete-x25h6\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.995981 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbb25\" (UniqueName: \"kubernetes.io/projected/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-kube-api-access-vbb25\") pod \"barbican92b5-account-delete-x25h6\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:03 crc kubenswrapper[4809]: I1205 11:34:03.997726 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-operator-scripts\") pod \"barbican92b5-account-delete-x25h6\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.997786 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.997792 4809 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.997834 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data podName:e1251531-cb99-418e-aa38-65a793e5a9d0 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:04.997816529 +0000 UTC m=+1540.388793167 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data") pod "rabbitmq-server-0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0") : configmap "rabbitmq-config-data" not found Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.997852 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:04.497843849 +0000 UTC m=+1539.888820527 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-config-data" not found Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.997885 4809 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 05 11:34:03 crc kubenswrapper[4809]: E1205 11:34:03.997903 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:04.497897951 +0000 UTC m=+1539.888874509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-scripts" not found Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.015210 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.056330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbb25\" (UniqueName: \"kubernetes.io/projected/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-kube-api-access-vbb25\") pod \"barbican92b5-account-delete-x25h6\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.089843 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi293e-account-delete-tb822"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.091578 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.093131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.101838 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.101928 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data podName:cfe522ab-449f-433e-a9bf-b33ed3b42595 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:04.601903184 +0000 UTC m=+1539.992879812 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data") pod "rabbitmq-cell1-server-0" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595") : configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.111844 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.121422 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi293e-account-delete-tb822"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.175185 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.204517 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts\") pod \"novaapi293e-account-delete-tb822\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.229806 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxqvc\" (UniqueName: \"kubernetes.io/projected/4ae41780-4bf1-4a48-bf11-74dadb59e387-kube-api-access-rxqvc\") pod \"novaapi293e-account-delete-tb822\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.213617 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-tcsq6"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.269106 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-tcsq6"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.283022 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell00f47-account-delete-qcdx8"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.293486 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.347782 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts\") pod \"novaapi293e-account-delete-tb822\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.347831 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxqvc\" (UniqueName: \"kubernetes.io/projected/4ae41780-4bf1-4a48-bf11-74dadb59e387-kube-api-access-rxqvc\") pod \"novaapi293e-account-delete-tb822\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.347880 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpt52\" (UniqueName: \"kubernetes.io/projected/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-kube-api-access-tpt52\") pod \"novacell00f47-account-delete-qcdx8\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.347921 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts\") pod \"novacell00f47-account-delete-qcdx8\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.348573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts\") pod \"novaapi293e-account-delete-tb822\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.350720 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell00f47-account-delete-qcdx8"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.377298 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2444g"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.386259 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxqvc\" (UniqueName: \"kubernetes.io/projected/4ae41780-4bf1-4a48-bf11-74dadb59e387-kube-api-access-rxqvc\") pod \"novaapi293e-account-delete-tb822\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.392706 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2444g"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.402005 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-bz8bw"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.413129 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fg4wz"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.449784 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.454138 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpt52\" (UniqueName: \"kubernetes.io/projected/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-kube-api-access-tpt52\") pod \"novacell00f47-account-delete-qcdx8\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.454297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts\") pod \"novacell00f47-account-delete-qcdx8\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.461109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts\") pod \"novacell00f47-account-delete-qcdx8\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.462736 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-94zqc"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.462980 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-94zqc" podUID="a5a2c6a1-9783-4742-948a-d992416fd6cc" containerName="openstack-network-exporter" containerID="cri-o://f41b7ad67c7562f28f0be7b6223e95bfb20a2e91b0355507be02a8e171f2fb04" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.494819 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ch8g6"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.495154 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" podUID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerName="dnsmasq-dns" containerID="cri-o://bd9495b849c3fa62b2b01b06cf243c2210d8b84430e1a712ed2a4c411f4baa90" gracePeriod=10 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.504958 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.505571 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-server" containerID="cri-o://5cdc142e3c8a66238160b2d52e5885ba4ddb6262d5ad5802d543fa708c536590" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.508364 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-updater" containerID="cri-o://213d6625154c146c9050604a6a30c02a271e04a3445369142a80d5fbd10042c8" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.508821 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="swift-recon-cron" containerID="cri-o://0414b93f70cfeebd7e27ab579af01c51e47ad9419cbb11d16f41369b2fe5cc96" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.508868 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="rsync" containerID="cri-o://e13456d6c296dda4598f34ee0b8c0e4703c954f9a8553fd24476fb75c023f44c" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.508901 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-expirer" containerID="cri-o://9d5533e9d925f25fad5f31fbd6f5b9fe99bee0d29f635ea849b663a235a46895" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.508933 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-updater" containerID="cri-o://640615ea9e04efceac48bfd84c7eaa5885da503591e80a899131fe2037fed294" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.508962 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-auditor" containerID="cri-o://22c2f8cd7c22d4a0109b044a886f55e278f8c205daf14ca5e1b47bd41b23426a" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509012 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-replicator" containerID="cri-o://6bb7e4f8fe910073653aa1a9f3a5a433f92b9f57a52d341885532226825ec2ca" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509041 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-server" containerID="cri-o://fc97466f1d8534b47cb24e449245a63275f1660daf521b08f83aaabeed242660" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509072 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-reaper" containerID="cri-o://c952b386502163d911d4615800b364440abeafbe8af4d4762f426ad2d8920b4c" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509102 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-auditor" containerID="cri-o://f644863166060fc77243c2059ca8902c14f11a6c9b85f76eb1d36cd09decb692" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509132 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-replicator" containerID="cri-o://d80d9b0fd1b858fe84af319a01ca8ae68a9151b6c070e69dff8050d858ebc3be" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509163 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-server" containerID="cri-o://58ee5b8b26614f934ab1ca4e41f1129f8f4b37fb56c414f5eef99f29744b4275" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509192 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-replicator" containerID="cri-o://2af085bac46d211c6f1357baf53c68b8c882c87ee5a214e91c8e24cfb34aaa57" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.509224 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-auditor" containerID="cri-o://0bb5aa4d3d56722e820bbab3a1408f0e2a8ad15ee9a7ade8db2984056113caab" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.527937 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpt52\" (UniqueName: \"kubernetes.io/projected/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-kube-api-access-tpt52\") pod \"novacell00f47-account-delete-qcdx8\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.543331 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-56b88bc7b6-8ckz6"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.544377 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-56b88bc7b6-8ckz6" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-log" containerID="cri-o://bf8cd41c788d38d0618e44a843797188ae9e6ab6ef244100c1e3de70041bbd8f" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.545077 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-56b88bc7b6-8ckz6" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-api" containerID="cri-o://eb470942b72b954581fcf3a32ef1e210fa14a93a6bab02c764a013a4e22b2998" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.556776 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cffd45dc7-dpz7j"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.560943 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cffd45dc7-dpz7j" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-api" containerID="cri-o://987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.561110 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cffd45dc7-dpz7j" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-httpd" containerID="cri-o://0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.562232 4809 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.562312 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:05.562266822 +0000 UTC m=+1540.953243380 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-config-data" not found Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.565140 4809 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.565208 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:05.565194777 +0000 UTC m=+1540.956171335 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-scripts" not found Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.593700 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.593974 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-log" containerID="cri-o://aa994132555d0692b38e19880ffec4d3ede1737f451a8f05d6f024366f509dfe" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.594130 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-httpd" containerID="cri-o://a82c49db9f73a1cf818cf2522001c715c552fa25f9dd1ad00a5785861ff76f4c" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.609369 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.609652 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="cinder-scheduler" containerID="cri-o://ed7d88f2530268a3f4438c5034a010adcd8f5b9f881c741aaee213b005a6a326" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.609793 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="probe" containerID="cri-o://76602c05e47629f8dd3a939796b77e7372f979cf7d26325edfe41b5b970ca71c" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.628678 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fqflm"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.648205 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.671668 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:04 crc kubenswrapper[4809]: E1205 11:34:04.671717 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data podName:cfe522ab-449f-433e-a9bf-b33ed3b42595 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:05.671705796 +0000 UTC m=+1541.062682354 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data") pod "rabbitmq-cell1-server-0" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595") : configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.697901 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fqflm"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.771670 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-qnwjc"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.832347 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-qnwjc"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.856295 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.856530 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-log" containerID="cri-o://35013bebb85eaa5b886089268483165db1836de375fb81c222beeadbb2a74fed" gracePeriod=30 Dec 05 11:34:04 crc kubenswrapper[4809]: I1205 11:34:04.858239 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-httpd" containerID="cri-o://635c738f9169739928f1ce8f18732784dac36a5c39b6e36889932cf1eda8f9e3" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:04.993020 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b7d8ff5-5828-4f49-9400-bdb47aa56419" path="/var/lib/kubelet/pods/0b7d8ff5-5828-4f49-9400-bdb47aa56419/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:04.993554 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64f0f74c-6a4b-4549-8abb-5d497627e1bf" path="/var/lib/kubelet/pods/64f0f74c-6a4b-4549-8abb-5d497627e1bf/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:04.994136 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ab6cf6-a8a1-43d5-8c37-8eddae072869" path="/var/lib/kubelet/pods/98ab6cf6-a8a1-43d5-8c37-8eddae072869/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.008601 4809 generic.go:334] "Generic (PLEG): container finished" podID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerID="bf8cd41c788d38d0618e44a843797188ae9e6ab6ef244100c1e3de70041bbd8f" exitCode=143 Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.021934 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a is running failed: container process not found" containerID="0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.030283 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac85b554-c4a8-483f-a64d-cedf0984140e" path="/var/lib/kubelet/pods/ac85b554-c4a8-483f-a64d-cedf0984140e/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.032205 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4e08826-2c05-4f4f-b899-7a87f44e51d5" path="/var/lib/kubelet/pods/b4e08826-2c05-4f4f-b899-7a87f44e51d5/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.034896 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae5c530-0398-48a0-ad06-2b08fd0d9c44" path="/var/lib/kubelet/pods/cae5c530-0398-48a0-ad06-2b08fd0d9c44/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.036592 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a is running failed: container process not found" containerID="0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.038548 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a is running failed: container process not found" containerID="0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.038598 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="ovsdbserver-nb" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.047317 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0e50e9-996a-4a27-aa6f-97bba30a0427" path="/var/lib/kubelet/pods/cc0e50e9-996a-4a27-aa6f-97bba30a0427/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.050840 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb75204b-f255-48c4-bc67-54bd3601c6a9" path="/var/lib/kubelet/pods/eb75204b-f255-48c4-bc67-54bd3601c6a9/volumes" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.051539 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.051563 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.051577 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b88bc7b6-8ckz6" event={"ID":"232c2be5-1e83-4fc2-b2c7-9a668d572d19","Type":"ContainerDied","Data":"bf8cd41c788d38d0618e44a843797188ae9e6ab6ef244100c1e3de70041bbd8f"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.067183 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementef3a-account-delete-qhc9d"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.086654 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.086712 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data podName:e1251531-cb99-418e-aa38-65a793e5a9d0 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:07.086694903 +0000 UTC m=+1542.477671461 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data") pod "rabbitmq-server-0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0") : configmap "rabbitmq-config-data" not found Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.092159 4809 generic.go:334] "Generic (PLEG): container finished" podID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerID="855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8" exitCode=2 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.092394 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3aabcdcf-4f51-4398-ae40-33b6d57aa514","Type":"ContainerDied","Data":"855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.094876 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.095102 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-log" containerID="cri-o://65c7a2783b79fc201d707bcdea31bba4d9626dc6a1edc4ed23ea125b15d958bf" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.095466 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-metadata" containerID="cri-o://6e8631852c5ee2d2ff12bc4e2bff28c710f77be32894780bc515ed6be7263ddd" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.105339 4809 generic.go:334] "Generic (PLEG): container finished" podID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerID="bd9495b849c3fa62b2b01b06cf243c2210d8b84430e1a712ed2a4c411f4baa90" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.105393 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" event={"ID":"462b7e18-c84b-42e8-ad9f-379bfe01bd26","Type":"ContainerDied","Data":"bd9495b849c3fa62b2b01b06cf243c2210d8b84430e1a712ed2a4c411f4baa90"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.146698 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.147102 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-log" containerID="cri-o://063a6dfa4af4111732f2aa27c4f351895860949b0c9d245dfe75cbda20b6e1cc" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.147373 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-api" containerID="cri-o://2e48dcfd66c9332d451a5a78ac5b9431515f09a61cddb083325ef0de295b6608" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169691 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="9d5533e9d925f25fad5f31fbd6f5b9fe99bee0d29f635ea849b663a235a46895" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169718 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="640615ea9e04efceac48bfd84c7eaa5885da503591e80a899131fe2037fed294" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169725 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="22c2f8cd7c22d4a0109b044a886f55e278f8c205daf14ca5e1b47bd41b23426a" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169731 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="213d6625154c146c9050604a6a30c02a271e04a3445369142a80d5fbd10042c8" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169737 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="f644863166060fc77243c2059ca8902c14f11a6c9b85f76eb1d36cd09decb692" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169743 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="d80d9b0fd1b858fe84af319a01ca8ae68a9151b6c070e69dff8050d858ebc3be" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169750 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="0bb5aa4d3d56722e820bbab3a1408f0e2a8ad15ee9a7ade8db2984056113caab" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169756 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="2af085bac46d211c6f1357baf53c68b8c882c87ee5a214e91c8e24cfb34aaa57" exitCode=0 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169796 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"9d5533e9d925f25fad5f31fbd6f5b9fe99bee0d29f635ea849b663a235a46895"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"640615ea9e04efceac48bfd84c7eaa5885da503591e80a899131fe2037fed294"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169829 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"22c2f8cd7c22d4a0109b044a886f55e278f8c205daf14ca5e1b47bd41b23426a"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"213d6625154c146c9050604a6a30c02a271e04a3445369142a80d5fbd10042c8"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169847 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"f644863166060fc77243c2059ca8902c14f11a6c9b85f76eb1d36cd09decb692"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"d80d9b0fd1b858fe84af319a01ca8ae68a9151b6c070e69dff8050d858ebc3be"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169865 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"0bb5aa4d3d56722e820bbab3a1408f0e2a8ad15ee9a7ade8db2984056113caab"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.169873 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"2af085bac46d211c6f1357baf53c68b8c882c87ee5a214e91c8e24cfb34aaa57"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.171814 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8434925e-b186-4558-956c-afa51f21113d/ovsdbserver-nb/0.log" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.171838 4809 generic.go:334] "Generic (PLEG): container finished" podID="8434925e-b186-4558-956c-afa51f21113d" containerID="17bf3535a2282730bf4584649cdef0fdc6c738363db31b9323fc6bba727833db" exitCode=2 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.171846 4809 generic.go:334] "Generic (PLEG): container finished" podID="8434925e-b186-4558-956c-afa51f21113d" containerID="0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a" exitCode=143 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.171874 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8434925e-b186-4558-956c-afa51f21113d","Type":"ContainerDied","Data":"17bf3535a2282730bf4584649cdef0fdc6c738363db31b9323fc6bba727833db"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.171888 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8434925e-b186-4558-956c-afa51f21113d","Type":"ContainerDied","Data":"0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.173289 4809 generic.go:334] "Generic (PLEG): container finished" podID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerID="aa994132555d0692b38e19880ffec4d3ede1737f451a8f05d6f024366f509dfe" exitCode=143 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.173324 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8bf5e0d0-86bc-4728-8ad4-678b4b28e190","Type":"ContainerDied","Data":"aa994132555d0692b38e19880ffec4d3ede1737f451a8f05d6f024366f509dfe"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.181845 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerName="rabbitmq" containerID="cri-o://aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30" gracePeriod=604800 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.182301 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-94zqc_a5a2c6a1-9783-4742-948a-d992416fd6cc/openstack-network-exporter/0.log" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.182332 4809 generic.go:334] "Generic (PLEG): container finished" podID="a5a2c6a1-9783-4742-948a-d992416fd6cc" containerID="f41b7ad67c7562f28f0be7b6223e95bfb20a2e91b0355507be02a8e171f2fb04" exitCode=2 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.182558 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api-log" containerID="cri-o://fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.182823 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-94zqc" event={"ID":"a5a2c6a1-9783-4742-948a-d992416fd6cc","Type":"ContainerDied","Data":"f41b7ad67c7562f28f0be7b6223e95bfb20a2e91b0355507be02a8e171f2fb04"} Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.184943 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api" containerID="cri-o://ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.185102 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.203231 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-68b55f4bc-9ngw8"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.208601 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-68b55f4bc-9ngw8" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker-log" containerID="cri-o://ffa1c61cecbaf03fa197f25a2a9928b22008f344ca29a0aacbc7194bddad9bc7" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.208824 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-68b55f4bc-9ngw8" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker" containerID="cri-o://929972b6f8b7d1846e06117ff7d6fe1f452d8c49bdedcc3dadd82bba1e6d2dc7" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.229686 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ktb2f"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.239006 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fd04fa46-c62b-4b8e-947d-da5846355feb/ovsdbserver-sb/0.log" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.239113 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.264028 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-69c49cc5f8-w57p4"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.264296 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener-log" containerID="cri-o://c9e36f881b5958ccc505a0d69b7688ee45bfcd1841d3987cabc75687e7c369c0" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.264406 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener" containerID="cri-o://bd369f5a80a27b3f543389e7e7b300b6bde366ca4736f6ed6dc27ce406afa252" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.295603 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-combined-ca-bundle\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.295662 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-config\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.295715 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbklf\" (UniqueName: \"kubernetes.io/projected/fd04fa46-c62b-4b8e-947d-da5846355feb-kube-api-access-fbklf\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.295891 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-metrics-certs-tls-certs\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.295939 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-scripts\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.295984 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.296001 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdbserver-sb-tls-certs\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.296081 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdb-rundir\") pod \"fd04fa46-c62b-4b8e-947d-da5846355feb\" (UID: \"fd04fa46-c62b-4b8e-947d-da5846355feb\") " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.298335 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-config" (OuterVolumeSpecName: "config") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.298671 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-scripts" (OuterVolumeSpecName: "scripts") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.299559 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.299960 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.299977 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.299989 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd04fa46-c62b-4b8e-947d-da5846355feb-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.329753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd04fa46-c62b-4b8e-947d-da5846355feb-kube-api-access-fbklf" (OuterVolumeSpecName: "kube-api-access-fbklf") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "kube-api-access-fbklf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.335851 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.367224 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ktb2f"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.412178 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.412213 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbklf\" (UniqueName: \"kubernetes.io/projected/fd04fa46-c62b-4b8e-947d-da5846355feb-kube-api-access-fbklf\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.422917 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-847649ff46-ldvld"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.423306 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-847649ff46-ldvld" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api-log" containerID="cri-o://9c770678dae206680aca70e609b6ceca94a601f7c7b778f6717417cdf5387456" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.423896 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-847649ff46-ldvld" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api" containerID="cri-o://f75988c31f2d6e2321aedeaffd4c2a8dcc59f31b1bcb9bc5bb4646d960e0a43b" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.462823 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.518592 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.529467 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.620551 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5fdb-account-create-update-92tdn"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.628029 4809 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.628065 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.628112 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:07.628091872 +0000 UTC m=+1543.019068430 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-scripts" not found Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.628207 4809 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.628263 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:07.628241415 +0000 UTC m=+1543.019217973 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-config-data" not found Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.653962 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.673649 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5fdb-account-create-update-92tdn"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.676278 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" containerID="cri-o://5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" gracePeriod=29 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.691057 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.698115 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "fd04fa46-c62b-4b8e-947d-da5846355feb" (UID: "fd04fa46-c62b-4b8e-947d-da5846355feb"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.707758 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.708014 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="50ec8106-d71a-45e3-af2d-f9bf27061233" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.729194 4809 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 05 11:34:05 crc kubenswrapper[4809]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 05 11:34:05 crc kubenswrapper[4809]: + source /usr/local/bin/container-scripts/functions Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNBridge=br-int Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNRemote=tcp:localhost:6642 Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNEncapType=geneve Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNAvailabilityZones= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ EnableChassisAsGateway=true Dec 05 11:34:05 crc kubenswrapper[4809]: ++ PhysicalNetworks= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNHostName= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 05 11:34:05 crc kubenswrapper[4809]: ++ ovs_dir=/var/lib/openvswitch Dec 05 11:34:05 crc kubenswrapper[4809]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 05 11:34:05 crc kubenswrapper[4809]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 05 11:34:05 crc kubenswrapper[4809]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + sleep 0.5 Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + sleep 0.5 Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + cleanup_ovsdb_server_semaphore Dec 05 11:34:05 crc kubenswrapper[4809]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 05 11:34:05 crc kubenswrapper[4809]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 05 11:34:05 crc kubenswrapper[4809]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-bz8bw" message=< Dec 05 11:34:05 crc kubenswrapper[4809]: Exiting ovsdb-server (5) [ OK ] Dec 05 11:34:05 crc kubenswrapper[4809]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 05 11:34:05 crc kubenswrapper[4809]: + source /usr/local/bin/container-scripts/functions Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNBridge=br-int Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNRemote=tcp:localhost:6642 Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNEncapType=geneve Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNAvailabilityZones= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ EnableChassisAsGateway=true Dec 05 11:34:05 crc kubenswrapper[4809]: ++ PhysicalNetworks= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNHostName= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 05 11:34:05 crc kubenswrapper[4809]: ++ ovs_dir=/var/lib/openvswitch Dec 05 11:34:05 crc kubenswrapper[4809]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 05 11:34:05 crc kubenswrapper[4809]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 05 11:34:05 crc kubenswrapper[4809]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + sleep 0.5 Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + sleep 0.5 Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + cleanup_ovsdb_server_semaphore Dec 05 11:34:05 crc kubenswrapper[4809]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 05 11:34:05 crc kubenswrapper[4809]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 05 11:34:05 crc kubenswrapper[4809]: > Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.729238 4809 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 05 11:34:05 crc kubenswrapper[4809]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 05 11:34:05 crc kubenswrapper[4809]: + source /usr/local/bin/container-scripts/functions Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNBridge=br-int Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNRemote=tcp:localhost:6642 Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNEncapType=geneve Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNAvailabilityZones= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ EnableChassisAsGateway=true Dec 05 11:34:05 crc kubenswrapper[4809]: ++ PhysicalNetworks= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ OVNHostName= Dec 05 11:34:05 crc kubenswrapper[4809]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 05 11:34:05 crc kubenswrapper[4809]: ++ ovs_dir=/var/lib/openvswitch Dec 05 11:34:05 crc kubenswrapper[4809]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 05 11:34:05 crc kubenswrapper[4809]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 05 11:34:05 crc kubenswrapper[4809]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + sleep 0.5 Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + sleep 0.5 Dec 05 11:34:05 crc kubenswrapper[4809]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 05 11:34:05 crc kubenswrapper[4809]: + cleanup_ovsdb_server_semaphore Dec 05 11:34:05 crc kubenswrapper[4809]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 05 11:34:05 crc kubenswrapper[4809]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 05 11:34:05 crc kubenswrapper[4809]: > pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" containerID="cri-o://911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.729279 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" containerID="cri-o://911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" gracePeriod=29 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.730761 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.730775 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd04fa46-c62b-4b8e-947d-da5846355feb-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.730831 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.730863 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data podName:cfe522ab-449f-433e-a9bf-b33ed3b42595 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:07.730851273 +0000 UTC m=+1543.121827841 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data") pod "rabbitmq-cell1-server-0" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595") : configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.743585 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.743772 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cc828bad-c058-4a1c-aa44-18d607759d6b" containerName="nova-scheduler-scheduler" containerID="cri-o://1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.750058 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerName="rabbitmq" containerID="cri-o://b1d76863d0e54906522b49358c89ce129152fe39aef470b4c3740799e8e8769d" gracePeriod=604800 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.757819 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.758016 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="23b3a18d-30f0-496f-9990-dd453ea8dbb7" containerName="nova-cell0-conductor-conductor" containerID="cri-o://57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.770693 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2pxck"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.795494 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2pxck"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.832107 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.832290 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="cd1f3174-82d6-4264-8e70-01f2ea709403" containerName="nova-cell1-conductor-conductor" containerID="cri-o://236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" gracePeriod=30 Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.839340 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l9n4x"] Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.865804 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l9n4x"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.877187 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.878071 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.879170 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:05 crc kubenswrapper[4809]: E1205 11:34:05.879196 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="cd1f3174-82d6-4264-8e70-01f2ea709403" containerName="nova-cell1-conductor-conductor" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.973279 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-94zqc_a5a2c6a1-9783-4742-948a-d992416fd6cc/openstack-network-exporter/0.log" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.974367 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:34:05 crc kubenswrapper[4809]: I1205 11:34:05.974526 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.038808 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-sb\") pod \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039062 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66227\" (UniqueName: \"kubernetes.io/projected/462b7e18-c84b-42e8-ad9f-379bfe01bd26-kube-api-access-66227\") pod \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039172 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-combined-ca-bundle\") pod \"a5a2c6a1-9783-4742-948a-d992416fd6cc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039239 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zv92\" (UniqueName: \"kubernetes.io/projected/a5a2c6a1-9783-4742-948a-d992416fd6cc-kube-api-access-9zv92\") pod \"a5a2c6a1-9783-4742-948a-d992416fd6cc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc\") pod \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039427 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovn-rundir\") pod \"a5a2c6a1-9783-4742-948a-d992416fd6cc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039489 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-config\") pod \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039559 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovs-rundir\") pod \"a5a2c6a1-9783-4742-948a-d992416fd6cc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039621 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-swift-storage-0\") pod \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039738 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-nb\") pod \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039807 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a2c6a1-9783-4742-948a-d992416fd6cc-config\") pod \"a5a2c6a1-9783-4742-948a-d992416fd6cc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.039891 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-metrics-certs-tls-certs\") pod \"a5a2c6a1-9783-4742-948a-d992416fd6cc\" (UID: \"a5a2c6a1-9783-4742-948a-d992416fd6cc\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.046058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "a5a2c6a1-9783-4742-948a-d992416fd6cc" (UID: "a5a2c6a1-9783-4742-948a-d992416fd6cc"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.051827 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.052221 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a5a2c6a1-9783-4742-948a-d992416fd6cc" (UID: "a5a2c6a1-9783-4742-948a-d992416fd6cc"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.053617 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5a2c6a1-9783-4742-948a-d992416fd6cc-config" (OuterVolumeSpecName: "config") pod "a5a2c6a1-9783-4742-948a-d992416fd6cc" (UID: "a5a2c6a1-9783-4742-948a-d992416fd6cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.060040 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.065985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/462b7e18-c84b-42e8-ad9f-379bfe01bd26-kube-api-access-66227" (OuterVolumeSpecName: "kube-api-access-66227") pod "462b7e18-c84b-42e8-ad9f-379bfe01bd26" (UID: "462b7e18-c84b-42e8-ad9f-379bfe01bd26"). InnerVolumeSpecName "kube-api-access-66227". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.073196 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.073256 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="ovn-northd" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.080598 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a2c6a1-9783-4742-948a-d992416fd6cc-kube-api-access-9zv92" (OuterVolumeSpecName: "kube-api-access-9zv92") pod "a5a2c6a1-9783-4742-948a-d992416fd6cc" (UID: "a5a2c6a1-9783-4742-948a-d992416fd6cc"). InnerVolumeSpecName "kube-api-access-9zv92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.110114 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance21cd-account-delete-kthxx"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.142920 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66227\" (UniqueName: \"kubernetes.io/projected/462b7e18-c84b-42e8-ad9f-379bfe01bd26-kube-api-access-66227\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.142948 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zv92\" (UniqueName: \"kubernetes.io/projected/a5a2c6a1-9783-4742-948a-d992416fd6cc-kube-api-access-9zv92\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.142958 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.142969 4809 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a5a2c6a1-9783-4742-948a-d992416fd6cc-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.142980 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5a2c6a1-9783-4742-948a-d992416fd6cc-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: W1205 11:34:06.183251 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6712dcd8_a4a4_4f57_adee_6b218823ba77.slice/crio-f2b2aeb6f62631fad0c5c779918f64fdaf8ad4b68a8440af51eb961ee601466c WatchSource:0}: Error finding container f2b2aeb6f62631fad0c5c779918f64fdaf8ad4b68a8440af51eb961ee601466c: Status 404 returned error can't find the container with id f2b2aeb6f62631fad0c5c779918f64fdaf8ad4b68a8440af51eb961ee601466c Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.195520 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-78555c6b85-lv8k2"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.195873 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-78555c6b85-lv8k2" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-httpd" containerID="cri-o://a3de22e1ff3f7898f9fee8f16d5df757205e16ec84e88652001b2d777b01c877" gracePeriod=30 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.196470 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-78555c6b85-lv8k2" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-server" containerID="cri-o://0c7d459938d65d4994155346009222c7ccaf5872fe4d76d2c6e4cb10a6a0e64d" gracePeriod=30 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.201089 4809 generic.go:334] "Generic (PLEG): container finished" podID="11752f85-2262-477b-9920-bf5c13f45f20" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.201155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bz8bw" event={"ID":"11752f85-2262-477b-9920-bf5c13f45f20","Type":"ContainerDied","Data":"911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.215204 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fd04fa46-c62b-4b8e-947d-da5846355feb/ovsdbserver-sb/0.log" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.215308 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fd04fa46-c62b-4b8e-947d-da5846355feb","Type":"ContainerDied","Data":"608ae4c80554477ca0ddaa85f9b68eadeb3d50c8750f03f07347c8727427fb4f"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.215342 4809 scope.go:117] "RemoveContainer" containerID="08a2f85174278c931e6dfac745c0fe3ac459b86ac3c72c047078366ce8f1b112" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.215483 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.243143 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerName="galera" containerID="cri-o://54582cab87c46307e3ea258499e465a65a3e2466f31df5e0ba4e63028feffc0f" gracePeriod=29 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.320578 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="e13456d6c296dda4598f34ee0b8c0e4703c954f9a8553fd24476fb75c023f44c" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.320603 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="6bb7e4f8fe910073653aa1a9f3a5a433f92b9f57a52d341885532226825ec2ca" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.320612 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="fc97466f1d8534b47cb24e449245a63275f1660daf521b08f83aaabeed242660" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.320621 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="58ee5b8b26614f934ab1ca4e41f1129f8f4b37fb56c414f5eef99f29744b4275" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.320627 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="c952b386502163d911d4615800b364440abeafbe8af4d4762f426ad2d8920b4c" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.320718 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="5cdc142e3c8a66238160b2d52e5885ba4ddb6262d5ad5802d543fa708c536590" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.321155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"e13456d6c296dda4598f34ee0b8c0e4703c954f9a8553fd24476fb75c023f44c"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.321186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"6bb7e4f8fe910073653aa1a9f3a5a433f92b9f57a52d341885532226825ec2ca"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.321196 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"fc97466f1d8534b47cb24e449245a63275f1660daf521b08f83aaabeed242660"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.321205 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"58ee5b8b26614f934ab1ca4e41f1129f8f4b37fb56c414f5eef99f29744b4275"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.321214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"c952b386502163d911d4615800b364440abeafbe8af4d4762f426ad2d8920b4c"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.321223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"5cdc142e3c8a66238160b2d52e5885ba4ddb6262d5ad5802d543fa708c536590"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.325076 4809 generic.go:334] "Generic (PLEG): container finished" podID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerID="fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2" exitCode=143 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.325131 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"846a6cff-41b5-4b49-8569-8c757669fe7f","Type":"ContainerDied","Data":"fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.325587 4809 scope.go:117] "RemoveContainer" containerID="f754ccd26270d79dad0c1628da75aaa14922d40a673887d3fffac5fd152c8685" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.329549 4809 generic.go:334] "Generic (PLEG): container finished" podID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" containerID="719fed1b9e606f1acf047af8258117088542a8ab9a3b33d882683eb029e590fd" exitCode=137 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.330387 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.331712 4809 generic.go:334] "Generic (PLEG): container finished" podID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerID="0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.332673 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffd45dc7-dpz7j" event={"ID":"b85b1950-6c23-424d-a724-a2a8cad9a05a","Type":"ContainerDied","Data":"0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.343564 4809 generic.go:334] "Generic (PLEG): container finished" podID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerID="35013bebb85eaa5b886089268483165db1836de375fb81c222beeadbb2a74fed" exitCode=143 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.343664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf8da287-1c2a-4d39-9d1b-dc5218b44907","Type":"ContainerDied","Data":"35013bebb85eaa5b886089268483165db1836de375fb81c222beeadbb2a74fed"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.346289 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.348151 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.348148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ch8g6" event={"ID":"462b7e18-c84b-42e8-ad9f-379bfe01bd26","Type":"ContainerDied","Data":"1656539d6fa79db56b38220e98941c55ca11a0efd1df637efdbbac04b8c6c519"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.352180 4809 generic.go:334] "Generic (PLEG): container finished" podID="871c924c-f749-40cd-81b2-64af559e6b20" containerID="65c7a2783b79fc201d707bcdea31bba4d9626dc6a1edc4ed23ea125b15d958bf" exitCode=143 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.352228 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"871c924c-f749-40cd-81b2-64af559e6b20","Type":"ContainerDied","Data":"65c7a2783b79fc201d707bcdea31bba4d9626dc6a1edc4ed23ea125b15d958bf"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.357828 4809 generic.go:334] "Generic (PLEG): container finished" podID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerID="929972b6f8b7d1846e06117ff7d6fe1f452d8c49bdedcc3dadd82bba1e6d2dc7" exitCode=0 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.357853 4809 generic.go:334] "Generic (PLEG): container finished" podID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerID="ffa1c61cecbaf03fa197f25a2a9928b22008f344ca29a0aacbc7194bddad9bc7" exitCode=143 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.357887 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b55f4bc-9ngw8" event={"ID":"95c323f7-b5da-4767-96dd-1c96f82c774d","Type":"ContainerDied","Data":"929972b6f8b7d1846e06117ff7d6fe1f452d8c49bdedcc3dadd82bba1e6d2dc7"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.357911 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b55f4bc-9ngw8" event={"ID":"95c323f7-b5da-4767-96dd-1c96f82c774d","Type":"ContainerDied","Data":"ffa1c61cecbaf03fa197f25a2a9928b22008f344ca29a0aacbc7194bddad9bc7"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.358604 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.360556 4809 generic.go:334] "Generic (PLEG): container finished" podID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerID="063a6dfa4af4111732f2aa27c4f351895860949b0c9d245dfe75cbda20b6e1cc" exitCode=143 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.360612 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8220cf04-d91a-45d8-ad53-51d0c6d21fe5","Type":"ContainerDied","Data":"063a6dfa4af4111732f2aa27c4f351895860949b0c9d245dfe75cbda20b6e1cc"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.367512 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5a2c6a1-9783-4742-948a-d992416fd6cc" (UID: "a5a2c6a1-9783-4742-948a-d992416fd6cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.373564 4809 generic.go:334] "Generic (PLEG): container finished" podID="919e76fa-7915-4a0f-8482-b827879dcacb" containerID="c9e36f881b5958ccc505a0d69b7688ee45bfcd1841d3987cabc75687e7c369c0" exitCode=143 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.373646 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" event={"ID":"919e76fa-7915-4a0f-8482-b827879dcacb","Type":"ContainerDied","Data":"c9e36f881b5958ccc505a0d69b7688ee45bfcd1841d3987cabc75687e7c369c0"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.374098 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.397095 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.403456 4809 generic.go:334] "Generic (PLEG): container finished" podID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerID="9c770678dae206680aca70e609b6ceca94a601f7c7b778f6717417cdf5387456" exitCode=143 Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.403577 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847649ff46-ldvld" event={"ID":"755354bd-175b-46e6-a9bf-93d8b9eab885","Type":"ContainerDied","Data":"9c770678dae206680aca70e609b6ceca94a601f7c7b778f6717417cdf5387456"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.420721 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-config" (OuterVolumeSpecName: "config") pod "462b7e18-c84b-42e8-ad9f-379bfe01bd26" (UID: "462b7e18-c84b-42e8-ad9f-379bfe01bd26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.427448 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.431202 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:06 crc kubenswrapper[4809]: E1205 11:34:06.431249 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="23b3a18d-30f0-496f-9990-dd453ea8dbb7" containerName="nova-cell0-conductor-conductor" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.439130 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementef3a-account-delete-qhc9d" event={"ID":"447a5847-564c-4b01-a74d-503249e0dd95","Type":"ContainerStarted","Data":"daeb3a766d910373c0287ca445155f5b98f2f4ac7e3057572652bfdc7767b0e2"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.439576 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "462b7e18-c84b-42e8-ad9f-379bfe01bd26" (UID: "462b7e18-c84b-42e8-ad9f-379bfe01bd26"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.443083 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8434925e-b186-4558-956c-afa51f21113d/ovsdbserver-nb/0.log" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.443153 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.449466 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "462b7e18-c84b-42e8-ad9f-379bfe01bd26" (UID: "462b7e18-c84b-42e8-ad9f-379bfe01bd26"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.449539 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-94zqc_a5a2c6a1-9783-4742-948a-d992416fd6cc/openstack-network-exporter/0.log" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.449658 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-94zqc" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.449743 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-94zqc" event={"ID":"a5a2c6a1-9783-4742-948a-d992416fd6cc","Type":"ContainerDied","Data":"bc480c77184046a20419642141ef662d97a30ddf1835770d77412fdb147cf956"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.449889 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-combined-ca-bundle\") pod \"95c323f7-b5da-4767-96dd-1c96f82c774d\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.449974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c323f7-b5da-4767-96dd-1c96f82c774d-logs\") pod \"95c323f7-b5da-4767-96dd-1c96f82c774d\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.449992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data-custom\") pod \"95c323f7-b5da-4767-96dd-1c96f82c774d\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450015 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqsln\" (UniqueName: \"kubernetes.io/projected/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-kube-api-access-hqsln\") pod \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config\") pod \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450102 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc\") pod \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\" (UID: \"462b7e18-c84b-42e8-ad9f-379bfe01bd26\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450186 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data\") pod \"95c323f7-b5da-4767-96dd-1c96f82c774d\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config-secret\") pod \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450266 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-combined-ca-bundle\") pod \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\" (UID: \"5a9a6752-88c3-4c2c-a54b-2e37a953fd50\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450330 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ttt9\" (UniqueName: \"kubernetes.io/projected/95c323f7-b5da-4767-96dd-1c96f82c774d-kube-api-access-4ttt9\") pod \"95c323f7-b5da-4767-96dd-1c96f82c774d\" (UID: \"95c323f7-b5da-4767-96dd-1c96f82c774d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450951 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450964 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.450975 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: W1205 11:34:06.452422 4809 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/462b7e18-c84b-42e8-ad9f-379bfe01bd26/volumes/kubernetes.io~configmap/dns-svc Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.452437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "462b7e18-c84b-42e8-ad9f-379bfe01bd26" (UID: "462b7e18-c84b-42e8-ad9f-379bfe01bd26"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.452608 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95c323f7-b5da-4767-96dd-1c96f82c774d-logs" (OuterVolumeSpecName: "logs") pod "95c323f7-b5da-4767-96dd-1c96f82c774d" (UID: "95c323f7-b5da-4767-96dd-1c96f82c774d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.454676 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "462b7e18-c84b-42e8-ad9f-379bfe01bd26" (UID: "462b7e18-c84b-42e8-ad9f-379bfe01bd26"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.454987 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "95c323f7-b5da-4767-96dd-1c96f82c774d" (UID: "95c323f7-b5da-4767-96dd-1c96f82c774d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.464217 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-kube-api-access-hqsln" (OuterVolumeSpecName: "kube-api-access-hqsln") pod "5a9a6752-88c3-4c2c-a54b-2e37a953fd50" (UID: "5a9a6752-88c3-4c2c-a54b-2e37a953fd50"). InnerVolumeSpecName "kube-api-access-hqsln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.468994 4809 scope.go:117] "RemoveContainer" containerID="bd9495b849c3fa62b2b01b06cf243c2210d8b84430e1a712ed2a4c411f4baa90" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.474584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance21cd-account-delete-kthxx" event={"ID":"6712dcd8-a4a4-4f57-adee-6b218823ba77","Type":"ContainerStarted","Data":"f2b2aeb6f62631fad0c5c779918f64fdaf8ad4b68a8440af51eb961ee601466c"} Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.475308 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placementef3a-account-delete-qhc9d" podStartSLOduration=3.47529497 podStartE2EDuration="3.47529497s" podCreationTimestamp="2025-12-05 11:34:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:34:06.464848331 +0000 UTC m=+1541.855824889" watchObservedRunningTime="2025-12-05 11:34:06.47529497 +0000 UTC m=+1541.866271518" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.487501 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95c323f7-b5da-4767-96dd-1c96f82c774d-kube-api-access-4ttt9" (OuterVolumeSpecName: "kube-api-access-4ttt9") pod "95c323f7-b5da-4767-96dd-1c96f82c774d" (UID: "95c323f7-b5da-4767-96dd-1c96f82c774d"). InnerVolumeSpecName "kube-api-access-4ttt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.529240 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a9a6752-88c3-4c2c-a54b-2e37a953fd50" (UID: "5a9a6752-88c3-4c2c-a54b-2e37a953fd50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.537547 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5a9a6752-88c3-4c2c-a54b-2e37a953fd50" (UID: "5a9a6752-88c3-4c2c-a54b-2e37a953fd50"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.539421 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a5a2c6a1-9783-4742-948a-d992416fd6cc" (UID: "a5a2c6a1-9783-4742-948a-d992416fd6cc"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.540995 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95c323f7-b5da-4767-96dd-1c96f82c774d" (UID: "95c323f7-b5da-4767-96dd-1c96f82c774d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-ovsdbserver-nb-tls-certs\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554426 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-config\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554456 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554567 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8434925e-b186-4558-956c-afa51f21113d-ovsdb-rundir\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554655 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-combined-ca-bundle\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554718 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqnk8\" (UniqueName: \"kubernetes.io/projected/8434925e-b186-4558-956c-afa51f21113d-kube-api-access-cqnk8\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554735 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-metrics-certs-tls-certs\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.554784 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-scripts\") pod \"8434925e-b186-4558-956c-afa51f21113d\" (UID: \"8434925e-b186-4558-956c-afa51f21113d\") " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555183 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a2c6a1-9783-4742-948a-d992416fd6cc-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555203 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555212 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555221 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ttt9\" (UniqueName: \"kubernetes.io/projected/95c323f7-b5da-4767-96dd-1c96f82c774d-kube-api-access-4ttt9\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555229 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555239 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c323f7-b5da-4767-96dd-1c96f82c774d-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555247 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555255 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqsln\" (UniqueName: \"kubernetes.io/projected/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-kube-api-access-hqsln\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555263 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.555272 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.559839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-config" (OuterVolumeSpecName: "config") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.562336 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8434925e-b186-4558-956c-afa51f21113d-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.563071 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-scripts" (OuterVolumeSpecName: "scripts") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.568627 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data" (OuterVolumeSpecName: "config-data") pod "95c323f7-b5da-4767-96dd-1c96f82c774d" (UID: "95c323f7-b5da-4767-96dd-1c96f82c774d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.568935 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.575293 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "462b7e18-c84b-42e8-ad9f-379bfe01bd26" (UID: "462b7e18-c84b-42e8-ad9f-379bfe01bd26"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.583053 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8434925e-b186-4558-956c-afa51f21113d-kube-api-access-cqnk8" (OuterVolumeSpecName: "kube-api-access-cqnk8") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "kube-api-access-cqnk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.597760 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5a9a6752-88c3-4c2c-a54b-2e37a953fd50" (UID: "5a9a6752-88c3-4c2c-a54b-2e37a953fd50"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.639265 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.659787 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqnk8\" (UniqueName: \"kubernetes.io/projected/8434925e-b186-4558-956c-afa51f21113d-kube-api-access-cqnk8\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.659992 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.660005 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8434925e-b186-4558-956c-afa51f21113d-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.660030 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.660040 4809 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/462b7e18-c84b-42e8-ad9f-379bfe01bd26-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.660152 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8434925e-b186-4558-956c-afa51f21113d-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.660162 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c323f7-b5da-4767-96dd-1c96f82c774d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.660171 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5a9a6752-88c3-4c2c-a54b-2e37a953fd50-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.660180 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.670909 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.714682 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.744089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican92b5-account-delete-x25h6"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.761104 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "8434925e-b186-4558-956c-afa51f21113d" (UID: "8434925e-b186-4558-956c-afa51f21113d"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.762342 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.762365 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8434925e-b186-4558-956c-afa51f21113d-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.762375 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.891436 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa2367d-dad7-461e-9e93-1ebc2bfee6e8" path="/var/lib/kubelet/pods/4aa2367d-dad7-461e-9e93-1ebc2bfee6e8/volumes" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.892029 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a9a6752-88c3-4c2c-a54b-2e37a953fd50" path="/var/lib/kubelet/pods/5a9a6752-88c3-4c2c-a54b-2e37a953fd50/volumes" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.892497 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad0cc4c-461f-4de2-aaf0-659c655edff3" path="/var/lib/kubelet/pods/5ad0cc4c-461f-4de2-aaf0-659c655edff3/volumes" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.894934 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73984ba-8f10-4cc7-aa62-30fa504ab55f" path="/var/lib/kubelet/pods/a73984ba-8f10-4cc7-aa62-30fa504ab55f/volumes" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.896980 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0813b74-3d43-48a9-a9af-738df2fe08ad" path="/var/lib/kubelet/pods/f0813b74-3d43-48a9-a9af-738df2fe08ad/volumes" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.897622 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" path="/var/lib/kubelet/pods/fd04fa46-c62b-4b8e-947d-da5846355feb/volumes" Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.898940 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder18b7-account-delete-jcsnx"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.945668 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi293e-account-delete-tb822"] Dec 05 11:34:06 crc kubenswrapper[4809]: I1205 11:34:06.959843 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronee61-account-delete-dfhwt"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.006213 4809 scope.go:117] "RemoveContainer" containerID="ceba1f0259ef6c16939487385f592f769bf0fd889b7aabff233d7301f28ef280" Dec 05 11:34:07 crc kubenswrapper[4809]: W1205 11:34:07.024134 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod102c9fa9_707d_4918_9a5d_55d8d4c47c12.slice/crio-c431b321f3820e7c612b4f190e35b1e4791949bc83b859b29302c641fa6b67aa WatchSource:0}: Error finding container c431b321f3820e7c612b4f190e35b1e4791949bc83b859b29302c641fa6b67aa: Status 404 returned error can't find the container with id c431b321f3820e7c612b4f190e35b1e4791949bc83b859b29302c641fa6b67aa Dec 05 11:34:07 crc kubenswrapper[4809]: W1205 11:34:07.039813 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ae41780_4bf1_4a48_bf11_74dadb59e387.slice/crio-dfef12a118ca5df7c5dba0bd8130f9595102f2bfeef282b56dce6ba5a62b7226 WatchSource:0}: Error finding container dfef12a118ca5df7c5dba0bd8130f9595102f2bfeef282b56dce6ba5a62b7226: Status 404 returned error can't find the container with id dfef12a118ca5df7c5dba0bd8130f9595102f2bfeef282b56dce6ba5a62b7226 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.118244 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell00f47-account-delete-qcdx8"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.152923 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.158180 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-94zqc"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.181362 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.181419 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data podName:e1251531-cb99-418e-aa38-65a793e5a9d0 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.181405207 +0000 UTC m=+1546.572381765 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data") pod "rabbitmq-server-0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0") : configmap "rabbitmq-config-data" not found Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.188863 4809 scope.go:117] "RemoveContainer" containerID="f41b7ad67c7562f28f0be7b6223e95bfb20a2e91b0355507be02a8e171f2fb04" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.189190 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-94zqc"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.201285 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ch8g6"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.208535 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ch8g6"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.222731 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.223236 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.223501 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.223624 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.223660 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.224586 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.228945 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.229018 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.283433 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-vencrypt-tls-certs\") pod \"50ec8106-d71a-45e3-af2d-f9bf27061233\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.283557 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-nova-novncproxy-tls-certs\") pod \"50ec8106-d71a-45e3-af2d-f9bf27061233\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.284588 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-combined-ca-bundle\") pod \"50ec8106-d71a-45e3-af2d-f9bf27061233\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.284723 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf8cr\" (UniqueName: \"kubernetes.io/projected/50ec8106-d71a-45e3-af2d-f9bf27061233-kube-api-access-mf8cr\") pod \"50ec8106-d71a-45e3-af2d-f9bf27061233\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.284897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-config-data\") pod \"50ec8106-d71a-45e3-af2d-f9bf27061233\" (UID: \"50ec8106-d71a-45e3-af2d-f9bf27061233\") " Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.294336 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50ec8106-d71a-45e3-af2d-f9bf27061233-kube-api-access-mf8cr" (OuterVolumeSpecName: "kube-api-access-mf8cr") pod "50ec8106-d71a-45e3-af2d-f9bf27061233" (UID: "50ec8106-d71a-45e3-af2d-f9bf27061233"). InnerVolumeSpecName "kube-api-access-mf8cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.333227 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50ec8106-d71a-45e3-af2d-f9bf27061233" (UID: "50ec8106-d71a-45e3-af2d-f9bf27061233"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.393828 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.393872 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf8cr\" (UniqueName: \"kubernetes.io/projected/50ec8106-d71a-45e3-af2d-f9bf27061233-kube-api-access-mf8cr\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.410866 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.411373 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-central-agent" containerID="cri-o://6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898" gracePeriod=30 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.411748 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="proxy-httpd" containerID="cri-o://36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292" gracePeriod=30 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.411782 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="sg-core" containerID="cri-o://d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509" gracePeriod=30 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.411767 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-notification-agent" containerID="cri-o://8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458" gracePeriod=30 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.475717 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.475976 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="21e68543-4956-4fc0-86aa-9ae8484b07c3" containerName="kube-state-metrics" containerID="cri-o://4770b253a022abba8402b06255199087a163db712ada152cea565a2565eb0ffc" gracePeriod=30 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.540360 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi293e-account-delete-tb822" event={"ID":"4ae41780-4bf1-4a48-bf11-74dadb59e387","Type":"ContainerStarted","Data":"dfef12a118ca5df7c5dba0bd8130f9595102f2bfeef282b56dce6ba5a62b7226"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.543174 4809 generic.go:334] "Generic (PLEG): container finished" podID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerID="0c7d459938d65d4994155346009222c7ccaf5872fe4d76d2c6e4cb10a6a0e64d" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.543203 4809 generic.go:334] "Generic (PLEG): container finished" podID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerID="a3de22e1ff3f7898f9fee8f16d5df757205e16ec84e88652001b2d777b01c877" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.543273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78555c6b85-lv8k2" event={"ID":"afdb0190-0a3d-482e-acbc-5d40e7999345","Type":"ContainerDied","Data":"0c7d459938d65d4994155346009222c7ccaf5872fe4d76d2c6e4cb10a6a0e64d"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.543303 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78555c6b85-lv8k2" event={"ID":"afdb0190-0a3d-482e-acbc-5d40e7999345","Type":"ContainerDied","Data":"a3de22e1ff3f7898f9fee8f16d5df757205e16ec84e88652001b2d777b01c877"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.543314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78555c6b85-lv8k2" event={"ID":"afdb0190-0a3d-482e-acbc-5d40e7999345","Type":"ContainerDied","Data":"c44c2c44654af38c08a5053cd17a9125148df8349fc3d004eff148bba529db6a"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.543323 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c44c2c44654af38c08a5053cd17a9125148df8349fc3d004eff148bba529db6a" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.548807 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-config-data" (OuterVolumeSpecName: "config-data") pod "50ec8106-d71a-45e3-af2d-f9bf27061233" (UID: "50ec8106-d71a-45e3-af2d-f9bf27061233"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.551322 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.551530 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="53c32262-ad33-4c11-9536-2b094ef77d55" containerName="memcached" containerID="cri-o://df1855ca03cbe3f85a7e21b167634d60144ff55554d113b302ffc0afc2b270ce" gracePeriod=30 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.560691 4809 generic.go:334] "Generic (PLEG): container finished" podID="447a5847-564c-4b01-a74d-503249e0dd95" containerID="9fcbf16b8475f5eccd9c0a78f2fdd8efeeeff59735de6888340ba66f89160e1f" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.560850 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementef3a-account-delete-qhc9d" event={"ID":"447a5847-564c-4b01-a74d-503249e0dd95","Type":"ContainerDied","Data":"9fcbf16b8475f5eccd9c0a78f2fdd8efeeeff59735de6888340ba66f89160e1f"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.589535 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8434925e-b186-4558-956c-afa51f21113d/ovsdbserver-nb/0.log" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.589707 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.590480 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8434925e-b186-4558-956c-afa51f21113d","Type":"ContainerDied","Data":"b0dd5ca362f30df1e4de05c60facfdc7242d5f815d785284025b685fb7a6da4a"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.590516 4809 scope.go:117] "RemoveContainer" containerID="17bf3535a2282730bf4584649cdef0fdc6c738363db31b9323fc6bba727833db" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.599348 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.608781 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="21e68543-4956-4fc0-86aa-9ae8484b07c3" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.169:8081/readyz\": dial tcp 10.217.0.169:8081: connect: connection refused" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.609111 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell00f47-account-delete-qcdx8" event={"ID":"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c","Type":"ContainerStarted","Data":"f1c13975c52f4b444845b4d98d685f5d9795f21536d4f0da75d936d87b8024ca"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.626107 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-77mnx"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.645128 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-77mnx"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.656988 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-22pc6"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.659701 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "50ec8106-d71a-45e3-af2d-f9bf27061233" (UID: "50ec8106-d71a-45e3-af2d-f9bf27061233"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.661096 4809 generic.go:334] "Generic (PLEG): container finished" podID="50ec8106-d71a-45e3-af2d-f9bf27061233" containerID="4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.661249 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50ec8106-d71a-45e3-af2d-f9bf27061233","Type":"ContainerDied","Data":"4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.661313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50ec8106-d71a-45e3-af2d-f9bf27061233","Type":"ContainerDied","Data":"212abebcd84168a4904d4dc0a7ee58aae12623d56536dec5af88f36cd1afa7d0"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.661421 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.663937 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-22pc6"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.670296 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone838e-account-delete-wv54s"] Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671525 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671552 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671571 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ec8106-d71a-45e3-af2d-f9bf27061233" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671579 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ec8106-d71a-45e3-af2d-f9bf27061233" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671591 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671597 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671611 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker-log" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671618 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker-log" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671775 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="ovsdbserver-sb" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671782 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="ovsdbserver-sb" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671795 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerName="init" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671800 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerName="init" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671810 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="ovsdbserver-nb" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671816 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="ovsdbserver-nb" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671827 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a2c6a1-9783-4742-948a-d992416fd6cc" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671832 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a2c6a1-9783-4742-948a-d992416fd6cc" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671843 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerName="dnsmasq-dns" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671850 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerName="dnsmasq-dns" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.671871 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.671877 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673598 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673646 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker-log" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673666 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673676 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" containerName="dnsmasq-dns" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673839 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd04fa46-c62b-4b8e-947d-da5846355feb" containerName="ovsdbserver-sb" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673852 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a2c6a1-9783-4742-948a-d992416fd6cc" containerName="openstack-network-exporter" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673870 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8434925e-b186-4558-956c-afa51f21113d" containerName="ovsdbserver-nb" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673877 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" containerName="barbican-worker" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.673890 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="50ec8106-d71a-45e3-af2d-f9bf27061233" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.675328 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.676018 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone838e-account-delete-wv54s"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.677256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b55f4bc-9ngw8" event={"ID":"95c323f7-b5da-4767-96dd-1c96f82c774d","Type":"ContainerDied","Data":"bc458e54ed56104cf41e973c977e2857cf8012cce3d3655c23cf3146f7117faa"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.683697 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-dbd7dc99f-5f57r"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.683984 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-dbd7dc99f-5f57r" podUID="f96a466a-b1fb-4260-b831-e4cd814a767e" containerName="keystone-api" containerID="cri-o://97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d" gracePeriod=30 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.685116 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b55f4bc-9ngw8" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.685865 4809 generic.go:334] "Generic (PLEG): container finished" podID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerID="54582cab87c46307e3ea258499e465a65a3e2466f31df5e0ba4e63028feffc0f" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.685928 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d2b5c2a4-1802-4308-9db8-090ea111f7ba","Type":"ContainerDied","Data":"54582cab87c46307e3ea258499e465a65a3e2466f31df5e0ba4e63028feffc0f"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.689617 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.700872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican92b5-account-delete-x25h6" event={"ID":"e409c78c-5de9-48ce-9c8b-afc67c6f8f48","Type":"ContainerStarted","Data":"03aa6fe669afc3afb549e6a085150e79d198307149b8f8ad7ff8518cad774a58"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.700924 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican92b5-account-delete-x25h6" event={"ID":"e409c78c-5de9-48ce-9c8b-afc67c6f8f48","Type":"ContainerStarted","Data":"beda354df71bd5e6d498a2e50405a359818fef035ea364934ca38f30750ceaa9"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.702412 4809 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.702527 4809 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.702581 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.702565323 +0000 UTC m=+1547.093541881 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-config-data" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.704150 4809 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.704231 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts podName:846a6cff-41b5-4b49-8569-8c757669fe7f nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.704211436 +0000 UTC m=+1547.095187994 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts") pod "cinder-api-0" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f") : secret "cinder-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.729570 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lxhkk"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.734077 4809 generic.go:334] "Generic (PLEG): container finished" podID="23b3a18d-30f0-496f-9990-dd453ea8dbb7" containerID="57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.734151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"23b3a18d-30f0-496f-9990-dd453ea8dbb7","Type":"ContainerDied","Data":"57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.743292 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lxhkk"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.758980 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.759863 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-838e-account-create-update-rt9bd"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.772205 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-838e-account-create-update-rt9bd"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.772438 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerID="76602c05e47629f8dd3a939796b77e7372f979cf7d26325edfe41b5b970ca71c" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.772498 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3d12a95a-49fa-4392-9a0d-9efd9aa28720","Type":"ContainerDied","Data":"76602c05e47629f8dd3a939796b77e7372f979cf7d26325edfe41b5b970ca71c"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.780228 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone838e-account-delete-wv54s"] Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.789683 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder18b7-account-delete-jcsnx" event={"ID":"102c9fa9-707d-4918-9a5d-55d8d4c47c12","Type":"ContainerStarted","Data":"3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.789728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder18b7-account-delete-jcsnx" event={"ID":"102c9fa9-707d-4918-9a5d-55d8d4c47c12","Type":"ContainerStarted","Data":"c431b321f3820e7c612b4f190e35b1e4791949bc83b859b29302c641fa6b67aa"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.790253 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder18b7-account-delete-jcsnx" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.798793 4809 generic.go:334] "Generic (PLEG): container finished" podID="6712dcd8-a4a4-4f57-adee-6b218823ba77" containerID="7db9bc6598923a51269cb0d779ffcff5d06cf7c742e40da04fae4c8d368b202b" exitCode=0 Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.798927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance21cd-account-delete-kthxx" event={"ID":"6712dcd8-a4a4-4f57-adee-6b218823ba77","Type":"ContainerDied","Data":"7db9bc6598923a51269cb0d779ffcff5d06cf7c742e40da04fae4c8d368b202b"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.800688 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronee61-account-delete-dfhwt" event={"ID":"77f68706-76e6-4ee3-b27d-bb092ab68a30","Type":"ContainerStarted","Data":"7e890444d0b95881a1015bb83540266f0364b4b9917cc820522d9cbe6336eece"} Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.801308 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutronee61-account-delete-dfhwt" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.808292 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.813255 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data podName:cfe522ab-449f-433e-a9bf-b33ed3b42595 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.813212678 +0000 UTC m=+1547.204189326 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data") pod "rabbitmq-cell1-server-0" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595") : configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.813316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.813542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gspz5\" (UniqueName: \"kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.814649 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder18b7-account-delete-jcsnx" podStartSLOduration=4.814614964 podStartE2EDuration="4.814614964s" podCreationTimestamp="2025-12-05 11:34:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:34:07.805187031 +0000 UTC m=+1543.196163599" watchObservedRunningTime="2025-12-05 11:34:07.814614964 +0000 UTC m=+1543.205591512" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.850441 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutronee61-account-delete-dfhwt" podStartSLOduration=4.850422838 podStartE2EDuration="4.850422838s" podCreationTimestamp="2025-12-05 11:34:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:34:07.833234365 +0000 UTC m=+1543.224210923" watchObservedRunningTime="2025-12-05 11:34:07.850422838 +0000 UTC m=+1543.241399386" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.878293 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "50ec8106-d71a-45e3-af2d-f9bf27061233" (UID: "50ec8106-d71a-45e3-af2d-f9bf27061233"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.919148 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.919518 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gspz5\" (UniqueName: \"kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:07 crc kubenswrapper[4809]: I1205 11:34:07.919611 4809 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ec8106-d71a-45e3-af2d-f9bf27061233-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.920535 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.920568 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts podName:102c9fa9-707d-4918-9a5d-55d8d4c47c12 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:08.420555758 +0000 UTC m=+1543.811532316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts") pod "cinder18b7-account-delete-jcsnx" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12") : configmap "openstack-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.921530 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.921556 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts podName:77e08e5f-18d4-4401-a19b-4852a792fa6e nodeName:}" failed. No retries permitted until 2025-12-05 11:34:08.421547693 +0000 UTC m=+1543.812524251 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts") pod "keystone838e-account-delete-wv54s" (UID: "77e08e5f-18d4-4401-a19b-4852a792fa6e") : configmap "openstack-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.921581 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.921600 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts podName:77f68706-76e6-4ee3-b27d-bb092ab68a30 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:08.421594464 +0000 UTC m=+1543.812571022 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts") pod "neutronee61-account-delete-dfhwt" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30") : configmap "openstack-scripts" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.925796 4809 projected.go:194] Error preparing data for projected volume kube-api-access-gspz5 for pod openstack/keystone838e-account-delete-wv54s: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 11:34:07 crc kubenswrapper[4809]: E1205 11:34:07.925860 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5 podName:77e08e5f-18d4-4401-a19b-4852a792fa6e nodeName:}" failed. No retries permitted until 2025-12-05 11:34:08.425842124 +0000 UTC m=+1543.816818682 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gspz5" (UniqueName: "kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5") pod "keystone838e-account-delete-wv54s" (UID: "77e08e5f-18d4-4401-a19b-4852a792fa6e") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.089996 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-56b88bc7b6-8ckz6" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.150:8778/\": read tcp 10.217.0.2:59128->10.217.0.150:8778: read: connection reset by peer" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.090099 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-56b88bc7b6-8ckz6" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.150:8778/\": read tcp 10.217.0.2:59132->10.217.0.150:8778: read: connection reset by peer" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.098057 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerName="galera" containerID="cri-o://cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f" gracePeriod=30 Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.106198 4809 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 05 11:34:08 crc kubenswrapper[4809]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-05T11:34:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 05 11:34:08 crc kubenswrapper[4809]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 05 11:34:08 crc kubenswrapper[4809]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-fg4wz" message=< Dec 05 11:34:08 crc kubenswrapper[4809]: Exiting ovn-controller (1) [FAILED] Dec 05 11:34:08 crc kubenswrapper[4809]: Killing ovn-controller (1) [ OK ] Dec 05 11:34:08 crc kubenswrapper[4809]: 2025-12-05T11:34:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 05 11:34:08 crc kubenswrapper[4809]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 05 11:34:08 crc kubenswrapper[4809]: > Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.106265 4809 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 05 11:34:08 crc kubenswrapper[4809]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-05T11:34:05Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 05 11:34:08 crc kubenswrapper[4809]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 05 11:34:08 crc kubenswrapper[4809]: > pod="openstack/ovn-controller-fg4wz" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" containerID="cri-o://6d5cb5ed862cbf7d2734d0573ce32c040186fb0fb53685e75922a348ec328b42" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.106307 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-fg4wz" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" containerID="cri-o://6d5cb5ed862cbf7d2734d0573ce32c040186fb0fb53685e75922a348ec328b42" gracePeriod=27 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.146489 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fg4wz" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" probeResult="failure" output="" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.211218 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-jllxv"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.223090 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-jllxv"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.231199 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementef3a-account-delete-qhc9d"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.266233 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ef3a-account-create-update-9mdwj"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.292168 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ef3a-account-create-update-9mdwj"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.356851 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-4ng99"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.370658 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-4ng99"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.413467 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.199:3000/\": dial tcp 10.217.0.199:3000: connect: connection refused" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.413815 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-21cd-account-create-update-hblhl"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.433114 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.433267 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gspz5\" (UniqueName: \"kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.433744 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.433811 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts podName:102c9fa9-707d-4918-9a5d-55d8d4c47c12 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:09.433793259 +0000 UTC m=+1544.824769817 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts") pod "cinder18b7-account-delete-jcsnx" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12") : configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.434202 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.434244 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts podName:77e08e5f-18d4-4401-a19b-4852a792fa6e nodeName:}" failed. No retries permitted until 2025-12-05 11:34:09.43423237 +0000 UTC m=+1544.825208938 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts") pod "keystone838e-account-delete-wv54s" (UID: "77e08e5f-18d4-4401-a19b-4852a792fa6e") : configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.434275 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.434297 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts podName:77f68706-76e6-4ee3-b27d-bb092ab68a30 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:09.434290572 +0000 UTC m=+1544.825267130 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts") pod "neutronee61-account-delete-dfhwt" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30") : configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.434506 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance21cd-account-delete-kthxx"] Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.437455 4809 projected.go:194] Error preparing data for projected volume kube-api-access-gspz5 for pod openstack/keystone838e-account-delete-wv54s: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.437545 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5 podName:77e08e5f-18d4-4401-a19b-4852a792fa6e nodeName:}" failed. No retries permitted until 2025-12-05 11:34:09.437519445 +0000 UTC m=+1544.828496063 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-gspz5" (UniqueName: "kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5") pod "keystone838e-account-delete-wv54s" (UID: "77e08e5f-18d4-4401-a19b-4852a792fa6e") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.444605 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-21cd-account-create-update-hblhl"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.549388 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.560388 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:33936->10.217.0.200:8775: read: connection reset by peer" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.560811 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:33938->10.217.0.200:8775: read: connection reset by peer" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.585029 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-86p6h"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.591567 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-86p6h"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.591682 4809 scope.go:117] "RemoveContainer" containerID="0ef0b17856f6be86763d75b0e8ec23e1335dfc559e33299c2e2a891845e30b4a" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.618916 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder18b7-account-delete-jcsnx"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.627226 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-18b7-account-create-update-99wvm"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.639477 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-18b7-account-create-update-99wvm"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653450 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-internal-tls-certs\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653506 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk7f6\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-kube-api-access-jk7f6\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-combined-ca-bundle\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-log-httpd\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653593 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-etc-swift\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653667 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-config-data\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653740 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-public-tls-certs\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.653806 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-run-httpd\") pod \"afdb0190-0a3d-482e-acbc-5d40e7999345\" (UID: \"afdb0190-0a3d-482e-acbc-5d40e7999345\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.654459 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.654471 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.669879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-kube-api-access-jk7f6" (OuterVolumeSpecName: "kube-api-access-jk7f6") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "kube-api-access-jk7f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.700570 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-2nbmc"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.700850 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.710556 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.717995 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-2nbmc"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.761878 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.761907 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk7f6\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-kube-api-access-jk7f6\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.761955 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afdb0190-0a3d-482e-acbc-5d40e7999345-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.761967 4809 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/afdb0190-0a3d-482e-acbc-5d40e7999345-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.766022 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.164:8776/healthcheck\": read tcp 10.217.0.2:35670->10.217.0.164:8776: read: connection reset by peer" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.778351 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ee61-account-create-update-d8glr"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.788882 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronee61-account-delete-dfhwt"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.795431 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ee61-account-create-update-d8glr"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.812050 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-bxdwk"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.820187 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-bxdwk"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.839198 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-92b5-account-create-update-8snn6"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.849451 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican92b5-account-delete-x25h6"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.853184 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.856048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.856876 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-92b5-account-create-update-8snn6"] Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.862956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"23b3a18d-30f0-496f-9990-dd453ea8dbb7","Type":"ContainerDied","Data":"c0a31ba61b268e35d82eeb2c4628dc61e5cfaa9fe1ee7bc86d26ed72479bcf4d"} Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.862993 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0a31ba61b268e35d82eeb2c4628dc61e5cfaa9fe1ee7bc86d26ed72479bcf4d" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863030 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863120 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-default\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45lnx\" (UniqueName: \"kubernetes.io/projected/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kube-api-access-45lnx\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863221 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-generated\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863249 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kolla-config\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863420 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-operator-scripts\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863451 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-galera-tls-certs\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863494 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-combined-ca-bundle\") pod \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\" (UID: \"d2b5c2a4-1802-4308-9db8-090ea111f7ba\") " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863956 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.863974 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.864274 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.865161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.866397 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.868481 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.873286 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kube-api-access-45lnx" (OuterVolumeSpecName: "kube-api-access-45lnx") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "kube-api-access-45lnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.873792 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fg4wz_e8226c08-75fc-4250-bc4d-e0dbbc818730/ovn-controller/0.log" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.873851 4809 generic.go:334] "Generic (PLEG): container finished" podID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerID="6d5cb5ed862cbf7d2734d0573ce32c040186fb0fb53685e75922a348ec328b42" exitCode=143 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.876163 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.877073 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell00f47-account-delete-qcdx8" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.882206 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.882933 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16be83eb-dd5e-4ddf-9d7d-ad6cba58a841" path="/var/lib/kubelet/pods/16be83eb-dd5e-4ddf-9d7d-ad6cba58a841/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.883768 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30d0bc12-98af-4fc5-bad7-509b69c92f79" path="/var/lib/kubelet/pods/30d0bc12-98af-4fc5-bad7-509b69c92f79/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.884324 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34702fbf-a89c-46ed-9546-bc873caee32e" path="/var/lib/kubelet/pods/34702fbf-a89c-46ed-9546-bc873caee32e/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.885309 4809 generic.go:334] "Generic (PLEG): container finished" podID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerID="a82c49db9f73a1cf818cf2522001c715c552fa25f9dd1ad00a5785861ff76f4c" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.885320 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a3f0d8a-08af-4640-b937-123d6e3405d7" path="/var/lib/kubelet/pods/3a3f0d8a-08af-4640-b937-123d6e3405d7/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.886388 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aede7af-428a-4d18-b78d-39e2c438d140" path="/var/lib/kubelet/pods/3aede7af-428a-4d18-b78d-39e2c438d140/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.887010 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="462b7e18-c84b-42e8-ad9f-379bfe01bd26" path="/var/lib/kubelet/pods/462b7e18-c84b-42e8-ad9f-379bfe01bd26/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.888512 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b2d9fda-bef0-4a73-84aa-c793181811d9" path="/var/lib/kubelet/pods/5b2d9fda-bef0-4a73-84aa-c793181811d9/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.890802 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1389c7-2aa9-4d9b-8070-a5e308a48ca2" path="/var/lib/kubelet/pods/8d1389c7-2aa9-4d9b-8070-a5e308a48ca2/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.891483 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutronee61-account-delete-dfhwt" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.892793 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-config-data" (OuterVolumeSpecName: "config-data") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.896199 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9868714a-454b-4647-9b49-58b91e834fb3" path="/var/lib/kubelet/pods/9868714a-454b-4647-9b49-58b91e834fb3/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.897066 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a2c6a1-9783-4742-948a-d992416fd6cc" path="/var/lib/kubelet/pods/a5a2c6a1-9783-4742-948a-d992416fd6cc/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.898120 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad966040-b1a2-448d-b673-aeada07d064f" path="/var/lib/kubelet/pods/ad966040-b1a2-448d-b673-aeada07d064f/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.898723 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afcf740a-2a63-438c-8cd9-774324492a8e" path="/var/lib/kubelet/pods/afcf740a-2a63-438c-8cd9-774324492a8e/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.900008 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c82cff53-0da9-4bed-baf0-4aa69f5c9074" path="/var/lib/kubelet/pods/c82cff53-0da9-4bed-baf0-4aa69f5c9074/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.901023 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced4cdcc-ec40-4177-a706-0386c12f1529" path="/var/lib/kubelet/pods/ced4cdcc-ec40-4177-a706-0386c12f1529/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.902549 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d074606c-4b4c-46bc-8552-d9a1ddcd9cc5" path="/var/lib/kubelet/pods/d074606c-4b4c-46bc-8552-d9a1ddcd9cc5/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.903896 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc29db8b-3c51-4b71-a984-41a1eb4a21c0" path="/var/lib/kubelet/pods/fc29db8b-3c51-4b71-a984-41a1eb4a21c0/volumes" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.905148 4809 generic.go:334] "Generic (PLEG): container finished" podID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerID="36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.905181 4809 generic.go:334] "Generic (PLEG): container finished" podID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerID="d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509" exitCode=2 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.905192 4809 generic.go:334] "Generic (PLEG): container finished" podID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerID="6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.905589 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.905663 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell00f47-account-delete-qcdx8" podStartSLOduration=4.905476368 podStartE2EDuration="4.905476368s" podCreationTimestamp="2025-12-05 11:34:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:34:08.895988784 +0000 UTC m=+1544.286965342" watchObservedRunningTime="2025-12-05 11:34:08.905476368 +0000 UTC m=+1544.296452926" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.909255 4809 generic.go:334] "Generic (PLEG): container finished" podID="e409c78c-5de9-48ce-9c8b-afc67c6f8f48" containerID="03aa6fe669afc3afb549e6a085150e79d198307149b8f8ad7ff8518cad774a58" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.913228 4809 generic.go:334] "Generic (PLEG): container finished" podID="21e68543-4956-4fc0-86aa-9ae8484b07c3" containerID="4770b253a022abba8402b06255199087a163db712ada152cea565a2565eb0ffc" exitCode=2 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.915990 4809 generic.go:334] "Generic (PLEG): container finished" podID="871c924c-f749-40cd-81b2-64af559e6b20" containerID="6e8631852c5ee2d2ff12bc4e2bff28c710f77be32894780bc515ed6be7263ddd" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.920295 4809 generic.go:334] "Generic (PLEG): container finished" podID="53c32262-ad33-4c11-9536-2b094ef77d55" containerID="df1855ca03cbe3f85a7e21b167634d60144ff55554d113b302ffc0afc2b270ce" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.922151 4809 generic.go:334] "Generic (PLEG): container finished" podID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerID="eb470942b72b954581fcf3a32ef1e210fa14a93a6bab02c764a013a4e22b2998" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.925200 4809 generic.go:334] "Generic (PLEG): container finished" podID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerID="635c738f9169739928f1ce8f18732784dac36a5c39b6e36889932cf1eda8f9e3" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.927455 4809 generic.go:334] "Generic (PLEG): container finished" podID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerID="2e48dcfd66c9332d451a5a78ac5b9431515f09a61cddb083325ef0de295b6608" exitCode=0 Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.928874 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78555c6b85-lv8k2" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.930479 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi293e-account-delete-tb822" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.931082 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder18b7-account-delete-jcsnx" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975822 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975852 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975873 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975883 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975893 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975903 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45lnx\" (UniqueName: \"kubernetes.io/projected/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kube-api-access-45lnx\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975915 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d2b5c2a4-1802-4308-9db8-090ea111f7ba-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.975926 4809 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d2b5c2a4-1802-4308-9db8-090ea111f7ba-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.984428 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: E1205 11:34:08.984486 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts podName:dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c nodeName:}" failed. No retries permitted until 2025-12-05 11:34:09.484469046 +0000 UTC m=+1544.875445604 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts") pod "novacell00f47-account-delete-qcdx8" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c") : configmap "openstack-scripts" not found Dec 05 11:34:08 crc kubenswrapper[4809]: I1205 11:34:08.997356 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "d2b5c2a4-1802-4308-9db8-090ea111f7ba" (UID: "d2b5c2a4-1802-4308-9db8-090ea111f7ba"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.047040 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.053848 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "afdb0190-0a3d-482e-acbc-5d40e7999345" (UID: "afdb0190-0a3d-482e-acbc-5d40e7999345"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.072528 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz" event={"ID":"e8226c08-75fc-4250-bc4d-e0dbbc818730","Type":"ContainerDied","Data":"6d5cb5ed862cbf7d2734d0573ce32c040186fb0fb53685e75922a348ec328b42"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.072575 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell00f47-account-delete-qcdx8" event={"ID":"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c","Type":"ContainerStarted","Data":"d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.072590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d2b5c2a4-1802-4308-9db8-090ea111f7ba","Type":"ContainerDied","Data":"f4e573e003cfef9d3f12853fee2d8c51a2d2a055035dfc75fef29ed7dcb19d04"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.072604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8bf5e0d0-86bc-4728-8ad4-678b4b28e190","Type":"ContainerDied","Data":"a82c49db9f73a1cf818cf2522001c715c552fa25f9dd1ad00a5785861ff76f4c"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronee61-account-delete-dfhwt" event={"ID":"77f68706-76e6-4ee3-b27d-bb092ab68a30","Type":"ContainerStarted","Data":"4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerDied","Data":"36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074095 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerDied","Data":"d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074104 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerDied","Data":"6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074113 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican92b5-account-delete-x25h6" event={"ID":"e409c78c-5de9-48ce-9c8b-afc67c6f8f48","Type":"ContainerDied","Data":"03aa6fe669afc3afb549e6a085150e79d198307149b8f8ad7ff8518cad774a58"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074126 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21e68543-4956-4fc0-86aa-9ae8484b07c3","Type":"ContainerDied","Data":"4770b253a022abba8402b06255199087a163db712ada152cea565a2565eb0ffc"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074138 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21e68543-4956-4fc0-86aa-9ae8484b07c3","Type":"ContainerDied","Data":"3233d6ed4f63a39d0cf7c9550d125b9e7c40bf22e15894ced65db71000e62605"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074151 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3233d6ed4f63a39d0cf7c9550d125b9e7c40bf22e15894ced65db71000e62605" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074163 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"871c924c-f749-40cd-81b2-64af559e6b20","Type":"ContainerDied","Data":"6e8631852c5ee2d2ff12bc4e2bff28c710f77be32894780bc515ed6be7263ddd"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"53c32262-ad33-4c11-9536-2b094ef77d55","Type":"ContainerDied","Data":"df1855ca03cbe3f85a7e21b167634d60144ff55554d113b302ffc0afc2b270ce"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074185 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b88bc7b6-8ckz6" event={"ID":"232c2be5-1e83-4fc2-b2c7-9a668d572d19","Type":"ContainerDied","Data":"eb470942b72b954581fcf3a32ef1e210fa14a93a6bab02c764a013a4e22b2998"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074198 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf8da287-1c2a-4d39-9d1b-dc5218b44907","Type":"ContainerDied","Data":"635c738f9169739928f1ce8f18732784dac36a5c39b6e36889932cf1eda8f9e3"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8220cf04-d91a-45d8-ad53-51d0c6d21fe5","Type":"ContainerDied","Data":"2e48dcfd66c9332d451a5a78ac5b9431515f09a61cddb083325ef0de295b6608"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.074222 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi293e-account-delete-tb822" event={"ID":"4ae41780-4bf1-4a48-bf11-74dadb59e387","Type":"ContainerStarted","Data":"1d84c3afeb7263e6079e6e47565f86fd4f1c3311f83150774d2707c65bc27b18"} Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.079702 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-gspz5 operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone838e-account-delete-wv54s" podUID="77e08e5f-18d4-4401-a19b-4852a792fa6e" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.084323 4809 scope.go:117] "RemoveContainer" containerID="4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.084970 4809 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2b5c2a4-1802-4308-9db8-090ea111f7ba-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.085006 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdb0190-0a3d-482e-acbc-5d40e7999345-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.085019 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.085487 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.085564 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts podName:4ae41780-4bf1-4a48-bf11-74dadb59e387 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:09.585542144 +0000 UTC m=+1544.976518702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts") pod "novaapi293e-account-delete-tb822" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387") : configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.087103 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi293e-account-delete-tb822" podStartSLOduration=5.087081004 podStartE2EDuration="5.087081004s" podCreationTimestamp="2025-12-05 11:34:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:34:08.990292807 +0000 UTC m=+1544.381269365" watchObservedRunningTime="2025-12-05 11:34:09.087081004 +0000 UTC m=+1544.478057552" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.092356 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-h8l7j"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.095571 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.115186 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-h8l7j"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.119343 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.133075 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-68b55f4bc-9ngw8"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.142824 4809 scope.go:117] "RemoveContainer" containerID="4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f" Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.144591 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f\": container with ID starting with 4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f not found: ID does not exist" containerID="4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.144658 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f"} err="failed to get container status \"4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f\": rpc error: code = NotFound desc = could not find container \"4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f\": container with ID starting with 4ddd34bdcb441af6bb0889f841be6aaabe5c461c1257f3da4d522b0a35c8f09f not found: ID does not exist" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.144684 4809 scope.go:117] "RemoveContainer" containerID="929972b6f8b7d1846e06117ff7d6fe1f452d8c49bdedcc3dadd82bba1e6d2dc7" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.149978 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-293e-account-create-update-82lc7"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.161864 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-68b55f4bc-9ngw8"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.169158 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi293e-account-delete-tb822"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.186219 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8n6k\" (UniqueName: \"kubernetes.io/projected/23b3a18d-30f0-496f-9990-dd453ea8dbb7-kube-api-access-p8n6k\") pod \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.186417 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-combined-ca-bundle\") pod \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.186739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-config-data\") pod \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\" (UID: \"23b3a18d-30f0-496f-9990-dd453ea8dbb7\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.191573 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23b3a18d-30f0-496f-9990-dd453ea8dbb7-kube-api-access-p8n6k" (OuterVolumeSpecName: "kube-api-access-p8n6k") pod "23b3a18d-30f0-496f-9990-dd453ea8dbb7" (UID: "23b3a18d-30f0-496f-9990-dd453ea8dbb7"). InnerVolumeSpecName "kube-api-access-p8n6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.233935 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.234693 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.234758 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-293e-account-create-update-82lc7"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.235182 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.237348 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.245555 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23b3a18d-30f0-496f-9990-dd453ea8dbb7" (UID: "23b3a18d-30f0-496f-9990-dd453ea8dbb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.259303 4809 scope.go:117] "RemoveContainer" containerID="ffa1c61cecbaf03fa197f25a2a9928b22008f344ca29a0aacbc7194bddad9bc7" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.282655 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.293866 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-certs\") pod \"21e68543-4956-4fc0-86aa-9ae8484b07c3\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.298008 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-combined-ca-bundle\") pod \"21e68543-4956-4fc0-86aa-9ae8484b07c3\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.298174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-config\") pod \"21e68543-4956-4fc0-86aa-9ae8484b07c3\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.298203 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62lp4\" (UniqueName: \"kubernetes.io/projected/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-api-access-62lp4\") pod \"21e68543-4956-4fc0-86aa-9ae8484b07c3\" (UID: \"21e68543-4956-4fc0-86aa-9ae8484b07c3\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.303809 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8n6k\" (UniqueName: \"kubernetes.io/projected/23b3a18d-30f0-496f-9990-dd453ea8dbb7-kube-api-access-p8n6k\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.303843 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.307170 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.308764 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-config-data" (OuterVolumeSpecName: "config-data") pod "23b3a18d-30f0-496f-9990-dd453ea8dbb7" (UID: "23b3a18d-30f0-496f-9990-dd453ea8dbb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.309248 4809 scope.go:117] "RemoveContainer" containerID="719fed1b9e606f1acf047af8258117088542a8ab9a3b33d882683eb029e590fd" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.317887 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.321016 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-api-access-62lp4" (OuterVolumeSpecName: "kube-api-access-62lp4") pod "21e68543-4956-4fc0-86aa-9ae8484b07c3" (UID: "21e68543-4956-4fc0-86aa-9ae8484b07c3"). InnerVolumeSpecName "kube-api-access-62lp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.322547 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.326933 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-847649ff46-ldvld" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": dial tcp 10.217.0.163:9311: connect: connection refused" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.334399 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6kb6p"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.338385 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6kb6p"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.347817 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-847649ff46-ldvld" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": dial tcp 10.217.0.163:9311: connect: connection refused" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.351602 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21e68543-4956-4fc0-86aa-9ae8484b07c3" (UID: "21e68543-4956-4fc0-86aa-9ae8484b07c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.356494 4809 scope.go:117] "RemoveContainer" containerID="54582cab87c46307e3ea258499e465a65a3e2466f31df5e0ba4e63028feffc0f" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.391352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "21e68543-4956-4fc0-86aa-9ae8484b07c3" (UID: "21e68543-4956-4fc0-86aa-9ae8484b07c3"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.392646 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "21e68543-4956-4fc0-86aa-9ae8484b07c3" (UID: "21e68543-4956-4fc0-86aa-9ae8484b07c3"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.404944 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-public-tls-certs\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405012 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-logs\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405099 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-combined-ca-bundle\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405171 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-internal-tls-certs\") pod \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405250 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqjbv\" (UniqueName: \"kubernetes.io/projected/232c2be5-1e83-4fc2-b2c7-9a668d572d19-kube-api-access-zqjbv\") pod \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-combined-ca-bundle\") pod \"53c32262-ad33-4c11-9536-2b094ef77d55\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405382 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-scripts\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-combined-ca-bundle\") pod \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405541 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-combined-ca-bundle\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405604 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-config-data\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405664 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhfzb\" (UniqueName: \"kubernetes.io/projected/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-kube-api-access-lhfzb\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405729 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-internal-tls-certs\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405773 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-config-data\") pod \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405802 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-httpd-run\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405837 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-config-data\") pod \"53c32262-ad33-4c11-9536-2b094ef77d55\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405870 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-config-data\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405904 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232c2be5-1e83-4fc2-b2c7-9a668d572d19-logs\") pod \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405956 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-scripts\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.405992 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-public-tls-certs\") pod \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406026 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-logs\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406048 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-httpd-run\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\" (UID: \"8bf5e0d0-86bc-4728-8ad4-678b4b28e190\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5x8m\" (UniqueName: \"kubernetes.io/projected/cf8da287-1c2a-4d39-9d1b-dc5218b44907-kube-api-access-w5x8m\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406215 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\" (UID: \"cf8da287-1c2a-4d39-9d1b-dc5218b44907\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406250 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-scripts\") pod \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\" (UID: \"232c2be5-1e83-4fc2-b2c7-9a668d572d19\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406284 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skkkn\" (UniqueName: \"kubernetes.io/projected/53c32262-ad33-4c11-9536-2b094ef77d55-kube-api-access-skkkn\") pod \"53c32262-ad33-4c11-9536-2b094ef77d55\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406337 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-memcached-tls-certs\") pod \"53c32262-ad33-4c11-9536-2b094ef77d55\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.406385 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-kolla-config\") pod \"53c32262-ad33-4c11-9536-2b094ef77d55\" (UID: \"53c32262-ad33-4c11-9536-2b094ef77d55\") " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407117 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407171 4809 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407194 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62lp4\" (UniqueName: \"kubernetes.io/projected/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-api-access-62lp4\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407209 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23b3a18d-30f0-496f-9990-dd453ea8dbb7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407221 4809 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e68543-4956-4fc0-86aa-9ae8484b07c3-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407254 4809 scope.go:117] "RemoveContainer" containerID="1f93774478558923bc3ec3b4f4d69eff36b9f1c081b7f2d5ab84f3d2a7b7c613" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407709 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.407887 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.408769 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "53c32262-ad33-4c11-9536-2b094ef77d55" (UID: "53c32262-ad33-4c11-9536-2b094ef77d55"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.409120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-config-data" (OuterVolumeSpecName: "config-data") pod "53c32262-ad33-4c11-9536-2b094ef77d55" (UID: "53c32262-ad33-4c11-9536-2b094ef77d55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.412499 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232c2be5-1e83-4fc2-b2c7-9a668d572d19-logs" (OuterVolumeSpecName: "logs") pod "232c2be5-1e83-4fc2-b2c7-9a668d572d19" (UID: "232c2be5-1e83-4fc2-b2c7-9a668d572d19"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.413305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.413753 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-logs" (OuterVolumeSpecName: "logs") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.414037 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf8da287-1c2a-4d39-9d1b-dc5218b44907-kube-api-access-w5x8m" (OuterVolumeSpecName: "kube-api-access-w5x8m") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "kube-api-access-w5x8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.415549 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.416123 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-logs" (OuterVolumeSpecName: "logs") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.418131 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-scripts" (OuterVolumeSpecName: "scripts") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.419380 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-scripts" (OuterVolumeSpecName: "scripts") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.421914 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c32262-ad33-4c11-9536-2b094ef77d55-kube-api-access-skkkn" (OuterVolumeSpecName: "kube-api-access-skkkn") pod "53c32262-ad33-4c11-9536-2b094ef77d55" (UID: "53c32262-ad33-4c11-9536-2b094ef77d55"). InnerVolumeSpecName "kube-api-access-skkkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.422625 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-scripts" (OuterVolumeSpecName: "scripts") pod "232c2be5-1e83-4fc2-b2c7-9a668d572d19" (UID: "232c2be5-1e83-4fc2-b2c7-9a668d572d19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.424616 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell00f47-account-delete-qcdx8"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.431283 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0f47-account-create-update-f7kxx"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.432039 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-kube-api-access-lhfzb" (OuterVolumeSpecName: "kube-api-access-lhfzb") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "kube-api-access-lhfzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.438984 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0f47-account-create-update-f7kxx"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.453167 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232c2be5-1e83-4fc2-b2c7-9a668d572d19-kube-api-access-zqjbv" (OuterVolumeSpecName: "kube-api-access-zqjbv") pod "232c2be5-1e83-4fc2-b2c7-9a668d572d19" (UID: "232c2be5-1e83-4fc2-b2c7-9a668d572d19"). InnerVolumeSpecName "kube-api-access-zqjbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.460536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53c32262-ad33-4c11-9536-2b094ef77d55" (UID: "53c32262-ad33-4c11-9536-2b094ef77d55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.485741 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.508815 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.508953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gspz5\" (UniqueName: \"kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5\") pod \"keystone838e-account-delete-wv54s\" (UID: \"77e08e5f-18d4-4401-a19b-4852a792fa6e\") " pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509058 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509078 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/232c2be5-1e83-4fc2-b2c7-9a668d572d19-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509091 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509099 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509107 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509131 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509140 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5x8m\" (UniqueName: \"kubernetes.io/projected/cf8da287-1c2a-4d39-9d1b-dc5218b44907-kube-api-access-w5x8m\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509156 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509165 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509175 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skkkn\" (UniqueName: \"kubernetes.io/projected/53c32262-ad33-4c11-9536-2b094ef77d55-kube-api-access-skkkn\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509183 4809 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/53c32262-ad33-4c11-9536-2b094ef77d55-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509192 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf8da287-1c2a-4d39-9d1b-dc5218b44907-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509203 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509213 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqjbv\" (UniqueName: \"kubernetes.io/projected/232c2be5-1e83-4fc2-b2c7-9a668d572d19-kube-api-access-zqjbv\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509227 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509238 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509251 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhfzb\" (UniqueName: \"kubernetes.io/projected/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-kube-api-access-lhfzb\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.509266 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510027 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510199 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts podName:77f68706-76e6-4ee3-b27d-bb092ab68a30 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.51017426 +0000 UTC m=+1546.901150878 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts") pod "neutronee61-account-delete-dfhwt" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30") : configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510311 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510390 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts podName:77e08e5f-18d4-4401-a19b-4852a792fa6e nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.510361105 +0000 UTC m=+1546.901337763 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts") pod "keystone838e-account-delete-wv54s" (UID: "77e08e5f-18d4-4401-a19b-4852a792fa6e") : configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510432 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510457 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts podName:102c9fa9-707d-4918-9a5d-55d8d4c47c12 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.510449177 +0000 UTC m=+1546.901425885 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts") pod "cinder18b7-account-delete-jcsnx" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12") : configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510728 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.510827 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts podName:dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c nodeName:}" failed. No retries permitted until 2025-12-05 11:34:10.510814536 +0000 UTC m=+1545.901791184 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts") pod "novacell00f47-account-delete-qcdx8" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c") : configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.521051 4809 projected.go:194] Error preparing data for projected volume kube-api-access-gspz5 for pod openstack/keystone838e-account-delete-wv54s: failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.521127 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5 podName:77e08e5f-18d4-4401-a19b-4852a792fa6e nodeName:}" failed. No retries permitted until 2025-12-05 11:34:11.521109042 +0000 UTC m=+1546.912085600 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-gspz5" (UniqueName: "kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5") pod "keystone838e-account-delete-wv54s" (UID: "77e08e5f-18d4-4401-a19b-4852a792fa6e") : failed to fetch token: serviceaccounts "galera-openstack" not found Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.532763 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.549751 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.555929 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.557546 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "232c2be5-1e83-4fc2-b2c7-9a668d572d19" (UID: "232c2be5-1e83-4fc2-b2c7-9a668d572d19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.564928 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-config-data" (OuterVolumeSpecName: "config-data") pod "232c2be5-1e83-4fc2-b2c7-9a668d572d19" (UID: "232c2be5-1e83-4fc2-b2c7-9a668d572d19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.565403 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-78555c6b85-lv8k2"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.572256 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-78555c6b85-lv8k2"] Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.586006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-config-data" (OuterVolumeSpecName: "config-data") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.593136 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.611305 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.611886 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.612004 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.612068 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.612133 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.612192 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.612298 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.612010 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: E1205 11:34:09.612586 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts podName:4ae41780-4bf1-4a48-bf11-74dadb59e387 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:10.612564461 +0000 UTC m=+1546.003541019 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts") pod "novaapi293e-account-delete-tb822" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387") : configmap "openstack-scripts" not found Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.613415 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.666006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "53c32262-ad33-4c11-9536-2b094ef77d55" (UID: "53c32262-ad33-4c11-9536-2b094ef77d55"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.668772 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-config-data" (OuterVolumeSpecName: "config-data") pod "8bf5e0d0-86bc-4728-8ad4-678b4b28e190" (UID: "8bf5e0d0-86bc-4728-8ad4-678b4b28e190"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.716316 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.718465 4809 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c32262-ad33-4c11-9536-2b094ef77d55-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.718602 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf5e0d0-86bc-4728-8ad4-678b4b28e190-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.745840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "232c2be5-1e83-4fc2-b2c7-9a668d572d19" (UID: "232c2be5-1e83-4fc2-b2c7-9a668d572d19"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.820417 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.859782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cf8da287-1c2a-4d39-9d1b-dc5218b44907" (UID: "cf8da287-1c2a-4d39-9d1b-dc5218b44907"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.882213 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "232c2be5-1e83-4fc2-b2c7-9a668d572d19" (UID: "232c2be5-1e83-4fc2-b2c7-9a668d572d19"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.922796 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf8da287-1c2a-4d39-9d1b-dc5218b44907-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.922835 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/232c2be5-1e83-4fc2-b2c7-9a668d572d19-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.942542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8bf5e0d0-86bc-4728-8ad4-678b4b28e190","Type":"ContainerDied","Data":"f0c1ffac9960b682a69cfc960ac40ad765fc8c81b9e4570edabb87578d4e2af5"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.942602 4809 scope.go:117] "RemoveContainer" containerID="a82c49db9f73a1cf818cf2522001c715c552fa25f9dd1ad00a5785861ff76f4c" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.942736 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.945372 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.948687 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.950099 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56b88bc7b6-8ckz6" event={"ID":"232c2be5-1e83-4fc2-b2c7-9a668d572d19","Type":"ContainerDied","Data":"d0170ca44a5ed898ac54a26ba30236a298a352d499632a3cc09f24a5f8a4bdf4"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.950175 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56b88bc7b6-8ckz6" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.952983 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cf8da287-1c2a-4d39-9d1b-dc5218b44907","Type":"ContainerDied","Data":"9c3771a79faa4fe599e47a439b6e595566c3c0558d1e7a143e2b85ca278d1f06"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.953149 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.953177 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.956377 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"871c924c-f749-40cd-81b2-64af559e6b20","Type":"ContainerDied","Data":"ecffc19de760f93ad003eed298aef231e8504b94d7049fff23cc61e0f2b3f211"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.956434 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.971982 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fg4wz_e8226c08-75fc-4250-bc4d-e0dbbc818730/ovn-controller/0.log" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.972073 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.977237 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.978656 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"53c32262-ad33-4c11-9536-2b094ef77d55","Type":"ContainerDied","Data":"c86e0256b321bd4d1de49795129a9fa46476f82cf92d0e6ae1dca7b1142a75a7"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.978744 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.983128 4809 scope.go:117] "RemoveContainer" containerID="aa994132555d0692b38e19880ffec4d3ede1737f451a8f05d6f024366f509dfe" Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.986987 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance21cd-account-delete-kthxx" event={"ID":"6712dcd8-a4a4-4f57-adee-6b218823ba77","Type":"ContainerDied","Data":"f2b2aeb6f62631fad0c5c779918f64fdaf8ad4b68a8440af51eb961ee601466c"} Dec 05 11:34:09 crc kubenswrapper[4809]: I1205 11:34:09.987017 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2b2aeb6f62631fad0c5c779918f64fdaf8ad4b68a8440af51eb961ee601466c" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.005237 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-fg4wz_e8226c08-75fc-4250-bc4d-e0dbbc818730/ovn-controller/0.log" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.005318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fg4wz" event={"ID":"e8226c08-75fc-4250-bc4d-e0dbbc818730","Type":"ContainerDied","Data":"8ef695dcb5b60d4deb4771eecf9462788c8a8b179c5543a1eb5405ec1f1e7ddf"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.005391 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fg4wz" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.005699 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.037134 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.038723 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.048445 4809 scope.go:117] "RemoveContainer" containerID="eb470942b72b954581fcf3a32ef1e210fa14a93a6bab02c764a013a4e22b2998" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.049755 4809 generic.go:334] "Generic (PLEG): container finished" podID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerID="8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458" exitCode=0 Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.049820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerDied","Data":"8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.049846 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39bb75c0-16a0-43d2-a344-aff2cd60b3c2","Type":"ContainerDied","Data":"ecc175d8ec4a837fb42632f9662bf5c71bc6bbb60432c51bf1729918933c1c30"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.061112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementef3a-account-delete-qhc9d" event={"ID":"447a5847-564c-4b01-a74d-503249e0dd95","Type":"ContainerDied","Data":"daeb3a766d910373c0287ca445155f5b98f2f4ac7e3057572652bfdc7767b0e2"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.061158 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="daeb3a766d910373c0287ca445155f5b98f2f4ac7e3057572652bfdc7767b0e2" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.061211 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementef3a-account-delete-qhc9d" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.067708 4809 generic.go:334] "Generic (PLEG): container finished" podID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerID="f75988c31f2d6e2321aedeaffd4c2a8dcc59f31b1bcb9bc5bb4646d960e0a43b" exitCode=0 Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.067779 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847649ff46-ldvld" event={"ID":"755354bd-175b-46e6-a9bf-93d8b9eab885","Type":"ContainerDied","Data":"f75988c31f2d6e2321aedeaffd4c2a8dcc59f31b1bcb9bc5bb4646d960e0a43b"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.072077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8220cf04-d91a-45d8-ad53-51d0c6d21fe5","Type":"ContainerDied","Data":"a22bd508ef19c64780c063829731f0660026372c8ade33583c9a9ecd7c2a2435"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.072183 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.081950 4809 generic.go:334] "Generic (PLEG): container finished" podID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerID="ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8" exitCode=0 Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.082025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"846a6cff-41b5-4b49-8569-8c757669fe7f","Type":"ContainerDied","Data":"ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.082055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"846a6cff-41b5-4b49-8569-8c757669fe7f","Type":"ContainerDied","Data":"7ed98d1e5c7dd98eea76484331349f5af670b8bd7119f23893508c5a1a79dc81"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.082130 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.086351 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican92b5-account-delete-x25h6" event={"ID":"e409c78c-5de9-48ce-9c8b-afc67c6f8f48","Type":"ContainerDied","Data":"beda354df71bd5e6d498a2e50405a359818fef035ea364934ca38f30750ceaa9"} Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.086390 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="beda354df71bd5e6d498a2e50405a359818fef035ea364934ca38f30750ceaa9" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.086384 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder18b7-account-delete-jcsnx" podUID="102c9fa9-707d-4918-9a5d-55d8d4c47c12" containerName="mariadb-account-delete" containerID="cri-o://3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73" gracePeriod=30 Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.086406 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.086453 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican92b5-account-delete-x25h6" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.086950 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell00f47-account-delete-qcdx8" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.087039 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.087074 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.087379 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutronee61-account-delete-dfhwt" podUID="77f68706-76e6-4ee3-b27d-bb092ab68a30" containerName="mariadb-account-delete" containerID="cri-o://4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8" gracePeriod=30 Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.088593 4809 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi293e-account-delete-tb822" secret="" err="secret \"galera-openstack-dockercfg-nj98s\" not found" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.123379 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.126449 4809 scope.go:117] "RemoveContainer" containerID="bf8cd41c788d38d0618e44a843797188ae9e6ab6ef244100c1e3de70041bbd8f" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138649 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871c924c-f749-40cd-81b2-64af559e6b20-logs\") pod \"871c924c-f749-40cd-81b2-64af559e6b20\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-combined-ca-bundle\") pod \"e8226c08-75fc-4250-bc4d-e0dbbc818730\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138730 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6712dcd8-a4a4-4f57-adee-6b218823ba77-operator-scripts\") pod \"6712dcd8-a4a4-4f57-adee-6b218823ba77\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138785 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138821 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run-ovn\") pod \"e8226c08-75fc-4250-bc4d-e0dbbc818730\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-operator-scripts\") pod \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8j5c\" (UniqueName: \"kubernetes.io/projected/e8226c08-75fc-4250-bc4d-e0dbbc818730-kube-api-access-p8j5c\") pod \"e8226c08-75fc-4250-bc4d-e0dbbc818730\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138902 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-config-data\") pod \"871c924c-f749-40cd-81b2-64af559e6b20\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138922 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-scripts\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run\") pod \"e8226c08-75fc-4250-bc4d-e0dbbc818730\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.138982 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-combined-ca-bundle\") pod \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139011 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cmhs\" (UniqueName: \"kubernetes.io/projected/846a6cff-41b5-4b49-8569-8c757669fe7f-kube-api-access-8cmhs\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8226c08-75fc-4250-bc4d-e0dbbc818730-scripts\") pod \"e8226c08-75fc-4250-bc4d-e0dbbc818730\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139068 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/447a5847-564c-4b01-a74d-503249e0dd95-operator-scripts\") pod \"447a5847-564c-4b01-a74d-503249e0dd95\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139110 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmbbv\" (UniqueName: \"kubernetes.io/projected/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-kube-api-access-kmbbv\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139136 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbb25\" (UniqueName: \"kubernetes.io/projected/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-kube-api-access-vbb25\") pod \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\" (UID: \"e409c78c-5de9-48ce-9c8b-afc67c6f8f48\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139154 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-sg-core-conf-yaml\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139178 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qftns\" (UniqueName: \"kubernetes.io/projected/6712dcd8-a4a4-4f57-adee-6b218823ba77-kube-api-access-qftns\") pod \"6712dcd8-a4a4-4f57-adee-6b218823ba77\" (UID: \"6712dcd8-a4a4-4f57-adee-6b218823ba77\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139208 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139231 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-log-httpd\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139260 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-config-data\") pod \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139284 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-public-tls-certs\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139306 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-combined-ca-bundle\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-config-data\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139357 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-logs\") pod \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139378 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-log-ovn\") pod \"e8226c08-75fc-4250-bc4d-e0dbbc818730\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139398 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2f29\" (UniqueName: \"kubernetes.io/projected/447a5847-564c-4b01-a74d-503249e0dd95-kube-api-access-n2f29\") pod \"447a5847-564c-4b01-a74d-503249e0dd95\" (UID: \"447a5847-564c-4b01-a74d-503249e0dd95\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139420 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-nova-metadata-tls-certs\") pod \"871c924c-f749-40cd-81b2-64af559e6b20\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139465 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-combined-ca-bundle\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139483 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-internal-tls-certs\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139510 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-ovn-controller-tls-certs\") pod \"e8226c08-75fc-4250-bc4d-e0dbbc818730\" (UID: \"e8226c08-75fc-4250-bc4d-e0dbbc818730\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139534 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-run-httpd\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139558 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-public-tls-certs\") pod \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846a6cff-41b5-4b49-8569-8c757669fe7f-logs\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139599 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/846a6cff-41b5-4b49-8569-8c757669fe7f-etc-machine-id\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-ceilometer-tls-certs\") pod \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\" (UID: \"39bb75c0-16a0-43d2-a344-aff2cd60b3c2\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139654 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-combined-ca-bundle\") pod \"871c924c-f749-40cd-81b2-64af559e6b20\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139678 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mnlp\" (UniqueName: \"kubernetes.io/projected/871c924c-f749-40cd-81b2-64af559e6b20-kube-api-access-5mnlp\") pod \"871c924c-f749-40cd-81b2-64af559e6b20\" (UID: \"871c924c-f749-40cd-81b2-64af559e6b20\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139702 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data-custom\") pod \"846a6cff-41b5-4b49-8569-8c757669fe7f\" (UID: \"846a6cff-41b5-4b49-8569-8c757669fe7f\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khlmm\" (UniqueName: \"kubernetes.io/projected/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-kube-api-access-khlmm\") pod \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.139752 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-internal-tls-certs\") pod \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\" (UID: \"8220cf04-d91a-45d8-ad53-51d0c6d21fe5\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.140735 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e8226c08-75fc-4250-bc4d-e0dbbc818730" (UID: "e8226c08-75fc-4250-bc4d-e0dbbc818730"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.140816 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/871c924c-f749-40cd-81b2-64af559e6b20-logs" (OuterVolumeSpecName: "logs") pod "871c924c-f749-40cd-81b2-64af559e6b20" (UID: "871c924c-f749-40cd-81b2-64af559e6b20"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.141837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e409c78c-5de9-48ce-9c8b-afc67c6f8f48" (UID: "e409c78c-5de9-48ce-9c8b-afc67c6f8f48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.145408 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6712dcd8-a4a4-4f57-adee-6b218823ba77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6712dcd8-a4a4-4f57-adee-6b218823ba77" (UID: "6712dcd8-a4a4-4f57-adee-6b218823ba77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.145731 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e8226c08-75fc-4250-bc4d-e0dbbc818730" (UID: "e8226c08-75fc-4250-bc4d-e0dbbc818730"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.146098 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-logs" (OuterVolumeSpecName: "logs") pod "8220cf04-d91a-45d8-ad53-51d0c6d21fe5" (UID: "8220cf04-d91a-45d8-ad53-51d0c6d21fe5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.146137 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/846a6cff-41b5-4b49-8569-8c757669fe7f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.148775 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run" (OuterVolumeSpecName: "var-run") pod "e8226c08-75fc-4250-bc4d-e0dbbc818730" (UID: "e8226c08-75fc-4250-bc4d-e0dbbc818730"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.149181 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8226c08-75fc-4250-bc4d-e0dbbc818730-scripts" (OuterVolumeSpecName: "scripts") pod "e8226c08-75fc-4250-bc4d-e0dbbc818730" (UID: "e8226c08-75fc-4250-bc4d-e0dbbc818730"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.151667 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.156391 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/447a5847-564c-4b01-a74d-503249e0dd95-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "447a5847-564c-4b01-a74d-503249e0dd95" (UID: "447a5847-564c-4b01-a74d-503249e0dd95"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.157135 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6712dcd8-a4a4-4f57-adee-6b218823ba77-kube-api-access-qftns" (OuterVolumeSpecName: "kube-api-access-qftns") pod "6712dcd8-a4a4-4f57-adee-6b218823ba77" (UID: "6712dcd8-a4a4-4f57-adee-6b218823ba77"). InnerVolumeSpecName "kube-api-access-qftns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.163867 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8226c08-75fc-4250-bc4d-e0dbbc818730-kube-api-access-p8j5c" (OuterVolumeSpecName: "kube-api-access-p8j5c") pod "e8226c08-75fc-4250-bc4d-e0dbbc818730" (UID: "e8226c08-75fc-4250-bc4d-e0dbbc818730"). InnerVolumeSpecName "kube-api-access-p8j5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.165879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-kube-api-access-kmbbv" (OuterVolumeSpecName: "kube-api-access-kmbbv") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "kube-api-access-kmbbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.169841 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.170779 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/447a5847-564c-4b01-a74d-503249e0dd95-kube-api-access-n2f29" (OuterVolumeSpecName: "kube-api-access-n2f29") pod "447a5847-564c-4b01-a74d-503249e0dd95" (UID: "447a5847-564c-4b01-a74d-503249e0dd95"). InnerVolumeSpecName "kube-api-access-n2f29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.171706 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/846a6cff-41b5-4b49-8569-8c757669fe7f-kube-api-access-8cmhs" (OuterVolumeSpecName: "kube-api-access-8cmhs") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "kube-api-access-8cmhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.172183 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/846a6cff-41b5-4b49-8569-8c757669fe7f-logs" (OuterVolumeSpecName: "logs") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.177741 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts" (OuterVolumeSpecName: "scripts") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.181271 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-scripts" (OuterVolumeSpecName: "scripts") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.186969 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-56b88bc7b6-8ckz6"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.187645 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-kube-api-access-vbb25" (OuterVolumeSpecName: "kube-api-access-vbb25") pod "e409c78c-5de9-48ce-9c8b-afc67c6f8f48" (UID: "e409c78c-5de9-48ce-9c8b-afc67c6f8f48"). InnerVolumeSpecName "kube-api-access-vbb25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.190189 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/871c924c-f749-40cd-81b2-64af559e6b20-kube-api-access-5mnlp" (OuterVolumeSpecName: "kube-api-access-5mnlp") pod "871c924c-f749-40cd-81b2-64af559e6b20" (UID: "871c924c-f749-40cd-81b2-64af559e6b20"). InnerVolumeSpecName "kube-api-access-5mnlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.204949 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.205235 4809 scope.go:117] "RemoveContainer" containerID="635c738f9169739928f1ce8f18732784dac36a5c39b6e36889932cf1eda8f9e3" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.210193 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.225489 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-kube-api-access-khlmm" (OuterVolumeSpecName: "kube-api-access-khlmm") pod "8220cf04-d91a-45d8-ad53-51d0c6d21fe5" (UID: "8220cf04-d91a-45d8-ad53-51d0c6d21fe5"). InnerVolumeSpecName "kube-api-access-khlmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.237626 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-config-data" (OuterVolumeSpecName: "config-data") pod "871c924c-f749-40cd-81b2-64af559e6b20" (UID: "871c924c-f749-40cd-81b2-64af559e6b20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.242963 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243011 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243024 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2f29\" (UniqueName: \"kubernetes.io/projected/447a5847-564c-4b01-a74d-503249e0dd95-kube-api-access-n2f29\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243033 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243042 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/846a6cff-41b5-4b49-8569-8c757669fe7f-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243054 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/846a6cff-41b5-4b49-8569-8c757669fe7f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243064 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mnlp\" (UniqueName: \"kubernetes.io/projected/871c924c-f749-40cd-81b2-64af559e6b20-kube-api-access-5mnlp\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243073 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243082 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khlmm\" (UniqueName: \"kubernetes.io/projected/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-kube-api-access-khlmm\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243094 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871c924c-f749-40cd-81b2-64af559e6b20-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243103 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6712dcd8-a4a4-4f57-adee-6b218823ba77-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243111 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243122 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243132 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243142 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8j5c\" (UniqueName: \"kubernetes.io/projected/e8226c08-75fc-4250-bc4d-e0dbbc818730-kube-api-access-p8j5c\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243154 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243170 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243180 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e8226c08-75fc-4250-bc4d-e0dbbc818730-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243189 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cmhs\" (UniqueName: \"kubernetes.io/projected/846a6cff-41b5-4b49-8569-8c757669fe7f-kube-api-access-8cmhs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243198 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8226c08-75fc-4250-bc4d-e0dbbc818730-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243209 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/447a5847-564c-4b01-a74d-503249e0dd95-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243218 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmbbv\" (UniqueName: \"kubernetes.io/projected/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-kube-api-access-kmbbv\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243229 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbb25\" (UniqueName: \"kubernetes.io/projected/e409c78c-5de9-48ce-9c8b-afc67c6f8f48-kube-api-access-vbb25\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243238 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qftns\" (UniqueName: \"kubernetes.io/projected/6712dcd8-a4a4-4f57-adee-6b218823ba77-kube-api-access-qftns\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.243250 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.266172 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-56b88bc7b6-8ckz6"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.278640 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8226c08-75fc-4250-bc4d-e0dbbc818730" (UID: "e8226c08-75fc-4250-bc4d-e0dbbc818730"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.284417 4809 scope.go:117] "RemoveContainer" containerID="35013bebb85eaa5b886089268483165db1836de375fb81c222beeadbb2a74fed" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.298778 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.303691 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.309405 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.317507 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.321610 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "871c924c-f749-40cd-81b2-64af559e6b20" (UID: "871c924c-f749-40cd-81b2-64af559e6b20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.327936 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.328686 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data" (OuterVolumeSpecName: "config-data") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.333344 4809 scope.go:117] "RemoveContainer" containerID="6e8631852c5ee2d2ff12bc4e2bff28c710f77be32894780bc515ed6be7263ddd" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.334175 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.344843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755354bd-175b-46e6-a9bf-93d8b9eab885-logs\") pod \"755354bd-175b-46e6-a9bf-93d8b9eab885\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.344934 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq692\" (UniqueName: \"kubernetes.io/projected/755354bd-175b-46e6-a9bf-93d8b9eab885-kube-api-access-bq692\") pod \"755354bd-175b-46e6-a9bf-93d8b9eab885\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.344965 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-internal-tls-certs\") pod \"755354bd-175b-46e6-a9bf-93d8b9eab885\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345097 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-combined-ca-bundle\") pod \"755354bd-175b-46e6-a9bf-93d8b9eab885\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345217 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-public-tls-certs\") pod \"755354bd-175b-46e6-a9bf-93d8b9eab885\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345302 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data\") pod \"755354bd-175b-46e6-a9bf-93d8b9eab885\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data-custom\") pod \"755354bd-175b-46e6-a9bf-93d8b9eab885\" (UID: \"755354bd-175b-46e6-a9bf-93d8b9eab885\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345761 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345791 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345802 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345811 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.345820 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.349298 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/755354bd-175b-46e6-a9bf-93d8b9eab885-logs" (OuterVolumeSpecName: "logs") pod "755354bd-175b-46e6-a9bf-93d8b9eab885" (UID: "755354bd-175b-46e6-a9bf-93d8b9eab885"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.352929 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.354841 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "755354bd-175b-46e6-a9bf-93d8b9eab885" (UID: "755354bd-175b-46e6-a9bf-93d8b9eab885"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.364803 4809 scope.go:117] "RemoveContainer" containerID="65c7a2783b79fc201d707bcdea31bba4d9626dc6a1edc4ed23ea125b15d958bf" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.364931 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.375911 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/755354bd-175b-46e6-a9bf-93d8b9eab885-kube-api-access-bq692" (OuterVolumeSpecName: "kube-api-access-bq692") pod "755354bd-175b-46e6-a9bf-93d8b9eab885" (UID: "755354bd-175b-46e6-a9bf-93d8b9eab885"). InnerVolumeSpecName "kube-api-access-bq692". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.375910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.385476 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8220cf04-d91a-45d8-ad53-51d0c6d21fe5" (UID: "8220cf04-d91a-45d8-ad53-51d0c6d21fe5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.397007 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.399603 4809 scope.go:117] "RemoveContainer" containerID="df1855ca03cbe3f85a7e21b167634d60144ff55554d113b302ffc0afc2b270ce" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.406242 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.411853 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.420120 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.422048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.426760 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementef3a-account-delete-qhc9d"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.435124 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementef3a-account-delete-qhc9d"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.442248 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican92b5-account-delete-x25h6"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.447517 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.447552 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.447562 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.447570 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755354bd-175b-46e6-a9bf-93d8b9eab885-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.447579 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq692\" (UniqueName: \"kubernetes.io/projected/755354bd-175b-46e6-a9bf-93d8b9eab885-kube-api-access-bq692\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.447590 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.449499 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican92b5-account-delete-x25h6"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.452116 4809 scope.go:117] "RemoveContainer" containerID="6d5cb5ed862cbf7d2734d0573ce32c040186fb0fb53685e75922a348ec328b42" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.465808 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-config-data" (OuterVolumeSpecName: "config-data") pod "8220cf04-d91a-45d8-ad53-51d0c6d21fe5" (UID: "8220cf04-d91a-45d8-ad53-51d0c6d21fe5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.475489 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "871c924c-f749-40cd-81b2-64af559e6b20" (UID: "871c924c-f749-40cd-81b2-64af559e6b20"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.493110 4809 scope.go:117] "RemoveContainer" containerID="36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.513098 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "755354bd-175b-46e6-a9bf-93d8b9eab885" (UID: "755354bd-175b-46e6-a9bf-93d8b9eab885"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.524821 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "846a6cff-41b5-4b49-8569-8c757669fe7f" (UID: "846a6cff-41b5-4b49-8569-8c757669fe7f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.535242 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "755354bd-175b-46e6-a9bf-93d8b9eab885" (UID: "755354bd-175b-46e6-a9bf-93d8b9eab885"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.548612 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.548650 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.548660 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/846a6cff-41b5-4b49-8569-8c757669fe7f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.548670 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/871c924c-f749-40cd-81b2-64af559e6b20-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.548679 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.548730 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.548740 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8220cf04-d91a-45d8-ad53-51d0c6d21fe5" (UID: "8220cf04-d91a-45d8-ad53-51d0c6d21fe5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.548797 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts podName:dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c nodeName:}" failed. No retries permitted until 2025-12-05 11:34:12.548779105 +0000 UTC m=+1547.939755663 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts") pod "novacell00f47-account-delete-qcdx8" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c") : configmap "openstack-scripts" not found Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.550064 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "e8226c08-75fc-4250-bc4d-e0dbbc818730" (UID: "e8226c08-75fc-4250-bc4d-e0dbbc818730"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.552276 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8220cf04-d91a-45d8-ad53-51d0c6d21fe5" (UID: "8220cf04-d91a-45d8-ad53-51d0c6d21fe5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.553796 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.555959 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-config-data" (OuterVolumeSpecName: "config-data") pod "39bb75c0-16a0-43d2-a344-aff2cd60b3c2" (UID: "39bb75c0-16a0-43d2-a344-aff2cd60b3c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.559793 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "755354bd-175b-46e6-a9bf-93d8b9eab885" (UID: "755354bd-175b-46e6-a9bf-93d8b9eab885"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.570171 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data" (OuterVolumeSpecName: "config-data") pod "755354bd-175b-46e6-a9bf-93d8b9eab885" (UID: "755354bd-175b-46e6-a9bf-93d8b9eab885"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.580400 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.581679 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.582700 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.582824 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cc828bad-c058-4a1c-aa44-18d607759d6b" containerName="nova-scheduler-scheduler" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.650333 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.650364 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39bb75c0-16a0-43d2-a344-aff2cd60b3c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.650373 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8226c08-75fc-4250-bc4d-e0dbbc818730-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.650383 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.650392 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.650400 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8220cf04-d91a-45d8-ad53-51d0c6d21fe5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.650408 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/755354bd-175b-46e6-a9bf-93d8b9eab885-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.650471 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.650517 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts podName:4ae41780-4bf1-4a48-bf11-74dadb59e387 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:12.6505031 +0000 UTC m=+1548.041479648 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts") pod "novaapi293e-account-delete-tb822" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387") : configmap "openstack-scripts" not found Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.666226 4809 scope.go:117] "RemoveContainer" containerID="d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.666614 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.682890 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.697848 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.703025 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fg4wz"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.709607 4809 scope.go:117] "RemoveContainer" containerID="8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.751543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-operator-scripts\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.751583 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kolla-config\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.751603 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-combined-ca-bundle\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.751979 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-galera-tls-certs\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.752017 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.752038 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-default\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.752135 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tncfv\" (UniqueName: \"kubernetes.io/projected/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kube-api-access-tncfv\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.752150 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-generated\") pod \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\" (UID: \"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9\") " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.753513 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.754316 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.755236 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.755808 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.763742 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kube-api-access-tncfv" (OuterVolumeSpecName: "kube-api-access-tncfv") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "kube-api-access-tncfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.771614 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fg4wz"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.798262 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.805891 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "mysql-db") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.808758 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" (UID: "6b760a19-f9d4-4ea7-a6c7-6610aa0334a9"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.811300 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.819444 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.830246 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.835931 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.860584 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.860754 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.860772 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.860787 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tncfv\" (UniqueName: \"kubernetes.io/projected/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kube-api-access-tncfv\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.861025 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.861975 4809 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.862016 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.862062 4809 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.876703 4809 scope.go:117] "RemoveContainer" containerID="6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.877802 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.880876 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.889752 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.892471 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e68543-4956-4fc0-86aa-9ae8484b07c3" path="/var/lib/kubelet/pods/21e68543-4956-4fc0-86aa-9ae8484b07c3/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.893297 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" path="/var/lib/kubelet/pods/232c2be5-1e83-4fc2-b2c7-9a668d572d19/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.893878 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23b3a18d-30f0-496f-9990-dd453ea8dbb7" path="/var/lib/kubelet/pods/23b3a18d-30f0-496f-9990-dd453ea8dbb7/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.894829 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2971236f-8f3a-4a90-bfcd-bfcdab8e1a24" path="/var/lib/kubelet/pods/2971236f-8f3a-4a90-bfcd-bfcdab8e1a24/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.895482 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b9c9e47-6fdc-4f5b-845f-503cf803036a" path="/var/lib/kubelet/pods/2b9c9e47-6fdc-4f5b-845f-503cf803036a/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.895470 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.896153 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="cd1f3174-82d6-4264-8e70-01f2ea709403" containerName="nova-cell1-conductor-conductor" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.897411 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="447a5847-564c-4b01-a74d-503249e0dd95" path="/var/lib/kubelet/pods/447a5847-564c-4b01-a74d-503249e0dd95/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.897902 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50ec8106-d71a-45e3-af2d-f9bf27061233" path="/var/lib/kubelet/pods/50ec8106-d71a-45e3-af2d-f9bf27061233/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.900716 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53c32262-ad33-4c11-9536-2b094ef77d55" path="/var/lib/kubelet/pods/53c32262-ad33-4c11-9536-2b094ef77d55/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.901200 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c316081-5125-4ba2-8a5d-87174b3066d2" path="/var/lib/kubelet/pods/6c316081-5125-4ba2-8a5d-87174b3066d2/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.901777 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" path="/var/lib/kubelet/pods/8220cf04-d91a-45d8-ad53-51d0c6d21fe5/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.903056 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8434925e-b186-4558-956c-afa51f21113d" path="/var/lib/kubelet/pods/8434925e-b186-4558-956c-afa51f21113d/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.903643 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" path="/var/lib/kubelet/pods/846a6cff-41b5-4b49-8569-8c757669fe7f/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.904212 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="871c924c-f749-40cd-81b2-64af559e6b20" path="/var/lib/kubelet/pods/871c924c-f749-40cd-81b2-64af559e6b20/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.905369 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" path="/var/lib/kubelet/pods/8bf5e0d0-86bc-4728-8ad4-678b4b28e190/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.906040 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95c323f7-b5da-4767-96dd-1c96f82c774d" path="/var/lib/kubelet/pods/95c323f7-b5da-4767-96dd-1c96f82c774d/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.907128 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" path="/var/lib/kubelet/pods/afdb0190-0a3d-482e-acbc-5d40e7999345/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.907657 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3931783-c543-4ec1-b91c-3b2a60c4b177" path="/var/lib/kubelet/pods/b3931783-c543-4ec1-b91c-3b2a60c4b177/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.908184 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" path="/var/lib/kubelet/pods/cf8da287-1c2a-4d39-9d1b-dc5218b44907/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.909275 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" path="/var/lib/kubelet/pods/d2b5c2a4-1802-4308-9db8-090ea111f7ba/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.909871 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e409c78c-5de9-48ce-9c8b-afc67c6f8f48" path="/var/lib/kubelet/pods/e409c78c-5de9-48ce-9c8b-afc67c6f8f48/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.910328 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" path="/var/lib/kubelet/pods/e8226c08-75fc-4250-bc4d-e0dbbc818730/volumes" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.913181 4809 scope.go:117] "RemoveContainer" containerID="36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.913674 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292\": container with ID starting with 36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292 not found: ID does not exist" containerID="36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.913737 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292"} err="failed to get container status \"36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292\": rpc error: code = NotFound desc = could not find container \"36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292\": container with ID starting with 36937ed97bbf564b68c715a5bbdc4cdb7170c10bd19b7b54ed8d1a33c5610292 not found: ID does not exist" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.913991 4809 scope.go:117] "RemoveContainer" containerID="d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.914675 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509\": container with ID starting with d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509 not found: ID does not exist" containerID="d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.914696 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509"} err="failed to get container status \"d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509\": rpc error: code = NotFound desc = could not find container \"d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509\": container with ID starting with d147dc3b7ada28bdfa8e646c6f128eb49d31a7956cafbd24d9e391abe595d509 not found: ID does not exist" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.914710 4809 scope.go:117] "RemoveContainer" containerID="8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.915069 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458\": container with ID starting with 8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458 not found: ID does not exist" containerID="8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.915141 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458"} err="failed to get container status \"8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458\": rpc error: code = NotFound desc = could not find container \"8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458\": container with ID starting with 8eebff86cd54978a34e03aafd5bbc6d03cddb1d72db4c3bcd1858a2f9f328458 not found: ID does not exist" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.915177 4809 scope.go:117] "RemoveContainer" containerID="6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898" Dec 05 11:34:10 crc kubenswrapper[4809]: E1205 11:34:10.915549 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898\": container with ID starting with 6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898 not found: ID does not exist" containerID="6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.915569 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898"} err="failed to get container status \"6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898\": rpc error: code = NotFound desc = could not find container \"6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898\": container with ID starting with 6627d3bc7b7ddd95dda9748c14e724ff0e1d3a5608806eaaac17ee0f9198d898 not found: ID does not exist" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.915582 4809 scope.go:117] "RemoveContainer" containerID="2e48dcfd66c9332d451a5a78ac5b9431515f09a61cddb083325ef0de295b6608" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.954588 4809 scope.go:117] "RemoveContainer" containerID="063a6dfa4af4111732f2aa27c4f351895860949b0c9d245dfe75cbda20b6e1cc" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.963816 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:10 crc kubenswrapper[4809]: I1205 11:34:10.984038 4809 scope.go:117] "RemoveContainer" containerID="ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.002705 4809 scope.go:117] "RemoveContainer" containerID="fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2" Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.006667 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.009054 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.010397 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.010474 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="ovn-northd" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.027795 4809 scope.go:117] "RemoveContainer" containerID="ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8" Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.028297 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8\": container with ID starting with ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8 not found: ID does not exist" containerID="ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.028348 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8"} err="failed to get container status \"ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8\": rpc error: code = NotFound desc = could not find container \"ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8\": container with ID starting with ab20557972c097fc9497187307e2580f23d3bbd14af2330dc65acfc764217be8 not found: ID does not exist" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.028419 4809 scope.go:117] "RemoveContainer" containerID="fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2" Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.028868 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2\": container with ID starting with fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2 not found: ID does not exist" containerID="fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.028906 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2"} err="failed to get container status \"fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2\": rpc error: code = NotFound desc = could not find container \"fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2\": container with ID starting with fc4429b6d65b3155c67456497bf93e8b58ed27408e858fcf203052b9f20f9cb2 not found: ID does not exist" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.112625 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847649ff46-ldvld" event={"ID":"755354bd-175b-46e6-a9bf-93d8b9eab885","Type":"ContainerDied","Data":"57f3fc6daee657204b26fd421cd7bad22af88f4ff2835b2839037129ffcda916"} Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.112675 4809 scope.go:117] "RemoveContainer" containerID="f75988c31f2d6e2321aedeaffd4c2a8dcc59f31b1bcb9bc5bb4646d960e0a43b" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.112768 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-847649ff46-ldvld" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.119616 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.124240 4809 generic.go:334] "Generic (PLEG): container finished" podID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerID="cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f" exitCode=0 Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.124294 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9","Type":"ContainerDied","Data":"cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f"} Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.124319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6b760a19-f9d4-4ea7-a6c7-6610aa0334a9","Type":"ContainerDied","Data":"805f4b013934c0f9ea85187e55df41599e25501b8b982044f1c99d18e7dc62e6"} Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.124296 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.130472 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell00f47-account-delete-qcdx8" podUID="dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" containerName="mariadb-account-delete" containerID="cri-o://d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18" gracePeriod=30 Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.130512 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance21cd-account-delete-kthxx" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.130476 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone838e-account-delete-wv54s" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.132561 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapi293e-account-delete-tb822" podUID="4ae41780-4bf1-4a48-bf11-74dadb59e387" containerName="mariadb-account-delete" containerID="cri-o://1d84c3afeb7263e6079e6e47565f86fd4f1c3311f83150774d2707c65bc27b18" gracePeriod=30 Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.142243 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-847649ff46-ldvld"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.148325 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/keystone-dbd7dc99f-5f57r" podUID="f96a466a-b1fb-4260-b831-e4cd814a767e" containerName="keystone-api" probeResult="failure" output="Get \"https://10.217.0.151:5000/v3\": read tcp 10.217.0.2:50960->10.217.0.151:5000: read: connection reset by peer" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.168820 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-847649ff46-ldvld"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.178083 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.178805 4809 scope.go:117] "RemoveContainer" containerID="9c770678dae206680aca70e609b6ceca94a601f7c7b778f6717417cdf5387456" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.182462 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.203047 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.212032 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.230353 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone838e-account-delete-wv54s"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.240707 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone838e-account-delete-wv54s"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.240730 4809 scope.go:117] "RemoveContainer" containerID="cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.244718 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance21cd-account-delete-kthxx"] Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.250129 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance21cd-account-delete-kthxx"] Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.280673 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.280752 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data podName:e1251531-cb99-418e-aa38-65a793e5a9d0 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:19.28073349 +0000 UTC m=+1554.671710048 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data") pod "rabbitmq-server-0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0") : configmap "rabbitmq-config-data" not found Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.308850 4809 scope.go:117] "RemoveContainer" containerID="2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.332739 4809 scope.go:117] "RemoveContainer" containerID="cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f" Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.333415 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f\": container with ID starting with cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f not found: ID does not exist" containerID="cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.333478 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f"} err="failed to get container status \"cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f\": rpc error: code = NotFound desc = could not find container \"cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f\": container with ID starting with cc129a18c4a0de87e89316c3005cce969e0f5368c0527b68d737e71f55c28b8f not found: ID does not exist" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.333512 4809 scope.go:117] "RemoveContainer" containerID="2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4" Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.333841 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4\": container with ID starting with 2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4 not found: ID does not exist" containerID="2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.333940 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4"} err="failed to get container status \"2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4\": rpc error: code = NotFound desc = could not find container \"2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4\": container with ID starting with 2da8bbdbd12c83509b53a07580e28b3c9586dc7f3ee84df158c1e63170b141c4 not found: ID does not exist" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.381783 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e08e5f-18d4-4401-a19b-4852a792fa6e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.381821 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gspz5\" (UniqueName: \"kubernetes.io/projected/77e08e5f-18d4-4401-a19b-4852a792fa6e-kube-api-access-gspz5\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.584902 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.585230 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts podName:77f68706-76e6-4ee3-b27d-bb092ab68a30 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:15.585213316 +0000 UTC m=+1550.976189894 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts") pod "neutronee61-account-delete-dfhwt" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30") : configmap "openstack-scripts" not found Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.585018 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.585383 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts podName:102c9fa9-707d-4918-9a5d-55d8d4c47c12 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:15.58536799 +0000 UTC m=+1550.976344548 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts") pod "cinder18b7-account-delete-jcsnx" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12") : configmap "openstack-scripts" not found Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.786427 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888314 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-confd\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888376 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e1251531-cb99-418e-aa38-65a793e5a9d0-erlang-cookie-secret\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888431 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-tls\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888464 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e1251531-cb99-418e-aa38-65a793e5a9d0-pod-info\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888491 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-plugins\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888537 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpxvp\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-kube-api-access-rpxvp\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888571 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888599 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-server-conf\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-plugins-conf\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-erlang-cookie\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.888837 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e1251531-cb99-418e-aa38-65a793e5a9d0\" (UID: \"e1251531-cb99-418e-aa38-65a793e5a9d0\") " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.889790 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.889967 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.890006 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.890091 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.890127 4809 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:11 crc kubenswrapper[4809]: E1205 11:34:11.890195 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data podName:cfe522ab-449f-433e-a9bf-b33ed3b42595 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:19.890171764 +0000 UTC m=+1555.281148422 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data") pod "rabbitmq-cell1-server-0" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595") : configmap "rabbitmq-cell1-config-data" not found Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.897057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.898646 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.899536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1251531-cb99-418e-aa38-65a793e5a9d0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.907942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e1251531-cb99-418e-aa38-65a793e5a9d0-pod-info" (OuterVolumeSpecName: "pod-info") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.913456 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-kube-api-access-rpxvp" (OuterVolumeSpecName: "kube-api-access-rpxvp") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "kube-api-access-rpxvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.995159 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpxvp\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-kube-api-access-rpxvp\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.995204 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.995243 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.995261 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e1251531-cb99-418e-aa38-65a793e5a9d0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.997137 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.997163 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e1251531-cb99-418e-aa38-65a793e5a9d0-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:11 crc kubenswrapper[4809]: I1205 11:34:11.997177 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.011894 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-server-conf" (OuterVolumeSpecName: "server-conf") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.016265 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.027923 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data" (OuterVolumeSpecName: "config-data") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.039562 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e1251531-cb99-418e-aa38-65a793e5a9d0" (UID: "e1251531-cb99-418e-aa38-65a793e5a9d0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.053509 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3aabcdcf-4f51-4398-ae40-33b6d57aa514/ovn-northd/0.log" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.053879 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.075071 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.099148 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.099179 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e1251531-cb99-418e-aa38-65a793e5a9d0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.099191 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.099201 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e1251531-cb99-418e-aa38-65a793e5a9d0-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.141842 4809 generic.go:334] "Generic (PLEG): container finished" podID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerID="b1d76863d0e54906522b49358c89ce129152fe39aef470b4c3740799e8e8769d" exitCode=0 Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.141908 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfe522ab-449f-433e-a9bf-b33ed3b42595","Type":"ContainerDied","Data":"b1d76863d0e54906522b49358c89ce129152fe39aef470b4c3740799e8e8769d"} Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.144845 4809 generic.go:334] "Generic (PLEG): container finished" podID="f96a466a-b1fb-4260-b831-e4cd814a767e" containerID="97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d" exitCode=0 Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.144929 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-dbd7dc99f-5f57r" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.144962 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-dbd7dc99f-5f57r" event={"ID":"f96a466a-b1fb-4260-b831-e4cd814a767e","Type":"ContainerDied","Data":"97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d"} Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.144997 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-dbd7dc99f-5f57r" event={"ID":"f96a466a-b1fb-4260-b831-e4cd814a767e","Type":"ContainerDied","Data":"c4a9049e8ab1203c1ef8f173799fcb79197f9b1e43843a9524df42f27268c5f3"} Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.145015 4809 scope.go:117] "RemoveContainer" containerID="97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.146933 4809 generic.go:334] "Generic (PLEG): container finished" podID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerID="aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30" exitCode=0 Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.146979 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e1251531-cb99-418e-aa38-65a793e5a9d0","Type":"ContainerDied","Data":"aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30"} Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.146998 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e1251531-cb99-418e-aa38-65a793e5a9d0","Type":"ContainerDied","Data":"fd93b120571515d66e8da7967dc33a5338e633e211a80c2962cea3e0498ea404"} Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.147044 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.158935 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3aabcdcf-4f51-4398-ae40-33b6d57aa514/ovn-northd/0.log" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.158976 4809 generic.go:334] "Generic (PLEG): container finished" podID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" exitCode=139 Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.159013 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3aabcdcf-4f51-4398-ae40-33b6d57aa514","Type":"ContainerDied","Data":"043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328"} Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.159034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3aabcdcf-4f51-4398-ae40-33b6d57aa514","Type":"ContainerDied","Data":"83d8e4cc27af30f9bc9fc36275cec53b7f0ff60bd01e0174cda8f42480d2b5bc"} Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.159080 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.172195 4809 scope.go:117] "RemoveContainer" containerID="97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.187891 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d\": container with ID starting with 97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d not found: ID does not exist" containerID="97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.187973 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d"} err="failed to get container status \"97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d\": rpc error: code = NotFound desc = could not find container \"97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d\": container with ID starting with 97e4e16b898799729f792475df42e4a7369a6b19956e90cb9b9fbfdc0ff4d22d not found: ID does not exist" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.188008 4809 scope.go:117] "RemoveContainer" containerID="aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.200489 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-combined-ca-bundle\") pod \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.200556 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqzdf\" (UniqueName: \"kubernetes.io/projected/f96a466a-b1fb-4260-b831-e4cd814a767e-kube-api-access-vqzdf\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.200590 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-config\") pod \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.200658 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-rundir\") pod \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.200705 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-metrics-certs-tls-certs\") pod \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.200737 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-scripts\") pod \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.200816 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-northd-tls-certs\") pod \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202297 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-combined-ca-bundle\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-scripts\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202374 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-internal-tls-certs\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-credential-keys\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202476 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbw6c\" (UniqueName: \"kubernetes.io/projected/3aabcdcf-4f51-4398-ae40-33b6d57aa514-kube-api-access-sbw6c\") pod \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\" (UID: \"3aabcdcf-4f51-4398-ae40-33b6d57aa514\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202538 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-public-tls-certs\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202581 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-config-data\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.202662 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-fernet-keys\") pod \"f96a466a-b1fb-4260-b831-e4cd814a767e\" (UID: \"f96a466a-b1fb-4260-b831-e4cd814a767e\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.213290 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-scripts" (OuterVolumeSpecName: "scripts") pod "3aabcdcf-4f51-4398-ae40-33b6d57aa514" (UID: "3aabcdcf-4f51-4398-ae40-33b6d57aa514"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.213494 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.214901 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-config" (OuterVolumeSpecName: "config") pod "3aabcdcf-4f51-4398-ae40-33b6d57aa514" (UID: "3aabcdcf-4f51-4398-ae40-33b6d57aa514"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.215597 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "3aabcdcf-4f51-4398-ae40-33b6d57aa514" (UID: "3aabcdcf-4f51-4398-ae40-33b6d57aa514"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.215705 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.220100 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.220322 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f96a466a-b1fb-4260-b831-e4cd814a767e-kube-api-access-vqzdf" (OuterVolumeSpecName: "kube-api-access-vqzdf") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "kube-api-access-vqzdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.223782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-scripts" (OuterVolumeSpecName: "scripts") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.223915 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.223960 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.224795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.225973 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.226023 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.234096 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.234921 4809 scope.go:117] "RemoveContainer" containerID="03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.234907 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.235018 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aabcdcf-4f51-4398-ae40-33b6d57aa514-kube-api-access-sbw6c" (OuterVolumeSpecName: "kube-api-access-sbw6c") pod "3aabcdcf-4f51-4398-ae40-33b6d57aa514" (UID: "3aabcdcf-4f51-4398-ae40-33b6d57aa514"). InnerVolumeSpecName "kube-api-access-sbw6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.247507 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.247575 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.250758 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.256848 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3aabcdcf-4f51-4398-ae40-33b6d57aa514" (UID: "3aabcdcf-4f51-4398-ae40-33b6d57aa514"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.258646 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-config-data" (OuterVolumeSpecName: "config-data") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.272099 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "3aabcdcf-4f51-4398-ae40-33b6d57aa514" (UID: "3aabcdcf-4f51-4398-ae40-33b6d57aa514"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.276443 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.283588 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f96a466a-b1fb-4260-b831-e4cd814a767e" (UID: "f96a466a-b1fb-4260-b831-e4cd814a767e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.285968 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "3aabcdcf-4f51-4398-ae40-33b6d57aa514" (UID: "3aabcdcf-4f51-4398-ae40-33b6d57aa514"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.289428 4809 scope.go:117] "RemoveContainer" containerID="aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.289946 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30\": container with ID starting with aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30 not found: ID does not exist" containerID="aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.289973 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30"} err="failed to get container status \"aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30\": rpc error: code = NotFound desc = could not find container \"aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30\": container with ID starting with aa9ed299a9c2e6e7f23cd3f38d074b198f385f96ba66314d75d2d8c6d8c84c30 not found: ID does not exist" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.289992 4809 scope.go:117] "RemoveContainer" containerID="03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.290313 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1\": container with ID starting with 03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1 not found: ID does not exist" containerID="03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.290331 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1"} err="failed to get container status \"03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1\": rpc error: code = NotFound desc = could not find container \"03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1\": container with ID starting with 03e548b4f3a53946745569c8c2a0b0b7f8b38f75ca77cb3c916445675e3755c1 not found: ID does not exist" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.290343 4809 scope.go:117] "RemoveContainer" containerID="855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307388 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307423 4809 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307433 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307442 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307454 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307467 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307476 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307485 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307493 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbw6c\" (UniqueName: \"kubernetes.io/projected/3aabcdcf-4f51-4398-ae40-33b6d57aa514-kube-api-access-sbw6c\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307502 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307509 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307519 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f96a466a-b1fb-4260-b831-e4cd814a767e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307527 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aabcdcf-4f51-4398-ae40-33b6d57aa514-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307535 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqzdf\" (UniqueName: \"kubernetes.io/projected/f96a466a-b1fb-4260-b831-e4cd814a767e-kube-api-access-vqzdf\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.307545 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aabcdcf-4f51-4398-ae40-33b6d57aa514-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.312189 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.314326 4809 scope.go:117] "RemoveContainer" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.353890 4809 scope.go:117] "RemoveContainer" containerID="855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.354356 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8\": container with ID starting with 855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8 not found: ID does not exist" containerID="855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.354415 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8"} err="failed to get container status \"855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8\": rpc error: code = NotFound desc = could not find container \"855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8\": container with ID starting with 855a96a02eff927eb350dd5fc8d394982512c3f9a985b4d2cb4596b368528ef8 not found: ID does not exist" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.354448 4809 scope.go:117] "RemoveContainer" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.354933 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328\": container with ID starting with 043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328 not found: ID does not exist" containerID="043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.354965 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328"} err="failed to get container status \"043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328\": rpc error: code = NotFound desc = could not find container \"043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328\": container with ID starting with 043bd9fd911c92390f66190be1fc3724797b906bd8e62392c06a9fcf73e23328 not found: ID does not exist" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409067 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-tls\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409129 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-confd\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409187 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409212 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqxrg\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-kube-api-access-sqxrg\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409236 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfe522ab-449f-433e-a9bf-b33ed3b42595-erlang-cookie-secret\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409300 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfe522ab-449f-433e-a9bf-b33ed3b42595-pod-info\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409334 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-server-conf\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409371 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-erlang-cookie\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-plugins\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409425 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-plugins-conf\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.409517 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cfe522ab-449f-433e-a9bf-b33ed3b42595\" (UID: \"cfe522ab-449f-433e-a9bf-b33ed3b42595\") " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.410989 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.411557 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.412058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.413151 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.413546 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/cfe522ab-449f-433e-a9bf-b33ed3b42595-pod-info" (OuterVolumeSpecName: "pod-info") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.414022 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe522ab-449f-433e-a9bf-b33ed3b42595-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.414043 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.414028 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-kube-api-access-sqxrg" (OuterVolumeSpecName: "kube-api-access-sqxrg") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "kube-api-access-sqxrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.430657 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data" (OuterVolumeSpecName: "config-data") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.452132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-server-conf" (OuterVolumeSpecName: "server-conf") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.493070 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-dbd7dc99f-5f57r"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.509312 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-dbd7dc99f-5f57r"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.511710 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.511898 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512056 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512178 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512322 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqxrg\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-kube-api-access-sqxrg\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512436 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfe522ab-449f-433e-a9bf-b33ed3b42595-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512541 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfe522ab-449f-433e-a9bf-b33ed3b42595-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512688 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfe522ab-449f-433e-a9bf-b33ed3b42595-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512802 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.512906 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.525941 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.528695 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.529443 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "cfe522ab-449f-433e-a9bf-b33ed3b42595" (UID: "cfe522ab-449f-433e-a9bf-b33ed3b42595"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.535290 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.614779 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.614985 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfe522ab-449f-433e-a9bf-b33ed3b42595-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.614928 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.615249 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts podName:dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c nodeName:}" failed. No retries permitted until 2025-12-05 11:34:16.615231431 +0000 UTC m=+1552.006207989 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts") pod "novacell00f47-account-delete-qcdx8" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c") : configmap "openstack-scripts" not found Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.716947 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:12 crc kubenswrapper[4809]: E1205 11:34:12.717129 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts podName:4ae41780-4bf1-4a48-bf11-74dadb59e387 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:16.717106929 +0000 UTC m=+1552.108083497 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts") pod "novaapi293e-account-delete-tb822" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387") : configmap "openstack-scripts" not found Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.889372 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" path="/var/lib/kubelet/pods/39bb75c0-16a0-43d2-a344-aff2cd60b3c2/volumes" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.890542 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" path="/var/lib/kubelet/pods/3aabcdcf-4f51-4398-ae40-33b6d57aa514/volumes" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.892040 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6712dcd8-a4a4-4f57-adee-6b218823ba77" path="/var/lib/kubelet/pods/6712dcd8-a4a4-4f57-adee-6b218823ba77/volumes" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.892886 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" path="/var/lib/kubelet/pods/6b760a19-f9d4-4ea7-a6c7-6610aa0334a9/volumes" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.893703 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" path="/var/lib/kubelet/pods/755354bd-175b-46e6-a9bf-93d8b9eab885/volumes" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.894977 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77e08e5f-18d4-4401-a19b-4852a792fa6e" path="/var/lib/kubelet/pods/77e08e5f-18d4-4401-a19b-4852a792fa6e/volumes" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.895661 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" path="/var/lib/kubelet/pods/e1251531-cb99-418e-aa38-65a793e5a9d0/volumes" Dec 05 11:34:12 crc kubenswrapper[4809]: I1205 11:34:12.896403 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f96a466a-b1fb-4260-b831-e4cd814a767e" path="/var/lib/kubelet/pods/f96a466a-b1fb-4260-b831-e4cd814a767e/volumes" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.176973 4809 generic.go:334] "Generic (PLEG): container finished" podID="cc828bad-c058-4a1c-aa44-18d607759d6b" containerID="1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2" exitCode=0 Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.177122 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc828bad-c058-4a1c-aa44-18d607759d6b","Type":"ContainerDied","Data":"1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2"} Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.177167 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cc828bad-c058-4a1c-aa44-18d607759d6b","Type":"ContainerDied","Data":"37dc18c66a331db4a85e69a97c224a913b4d46939aab7b1a450f853642d114fe"} Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.177181 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37dc18c66a331db4a85e69a97c224a913b4d46939aab7b1a450f853642d114fe" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.183061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfe522ab-449f-433e-a9bf-b33ed3b42595","Type":"ContainerDied","Data":"63a74e3b8be4d2e13bf4283513c5308ca14b97ff351ef8e5226ff822788891ac"} Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.183100 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.183109 4809 scope.go:117] "RemoveContainer" containerID="b1d76863d0e54906522b49358c89ce129152fe39aef470b4c3740799e8e8769d" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.186958 4809 generic.go:334] "Generic (PLEG): container finished" podID="919e76fa-7915-4a0f-8482-b827879dcacb" containerID="bd369f5a80a27b3f543389e7e7b300b6bde366ca4736f6ed6dc27ce406afa252" exitCode=0 Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.187003 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" event={"ID":"919e76fa-7915-4a0f-8482-b827879dcacb","Type":"ContainerDied","Data":"bd369f5a80a27b3f543389e7e7b300b6bde366ca4736f6ed6dc27ce406afa252"} Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.211831 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.227530 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.238580 4809 scope.go:117] "RemoveContainer" containerID="406129ebd40a9895b56eb4df3cb21847834b08c83a4aab01708e82905b319f2a" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.267069 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.325803 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwlkw\" (UniqueName: \"kubernetes.io/projected/cc828bad-c058-4a1c-aa44-18d607759d6b-kube-api-access-dwlkw\") pod \"cc828bad-c058-4a1c-aa44-18d607759d6b\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.326077 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-combined-ca-bundle\") pod \"cc828bad-c058-4a1c-aa44-18d607759d6b\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.326170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-config-data\") pod \"cc828bad-c058-4a1c-aa44-18d607759d6b\" (UID: \"cc828bad-c058-4a1c-aa44-18d607759d6b\") " Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.330812 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc828bad-c058-4a1c-aa44-18d607759d6b-kube-api-access-dwlkw" (OuterVolumeSpecName: "kube-api-access-dwlkw") pod "cc828bad-c058-4a1c-aa44-18d607759d6b" (UID: "cc828bad-c058-4a1c-aa44-18d607759d6b"). InnerVolumeSpecName "kube-api-access-dwlkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.346200 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-config-data" (OuterVolumeSpecName: "config-data") pod "cc828bad-c058-4a1c-aa44-18d607759d6b" (UID: "cc828bad-c058-4a1c-aa44-18d607759d6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.351092 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc828bad-c058-4a1c-aa44-18d607759d6b" (UID: "cc828bad-c058-4a1c-aa44-18d607759d6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.427603 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.427659 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwlkw\" (UniqueName: \"kubernetes.io/projected/cc828bad-c058-4a1c-aa44-18d607759d6b-kube-api-access-dwlkw\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.427675 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc828bad-c058-4a1c-aa44-18d607759d6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:13 crc kubenswrapper[4809]: I1205 11:34:13.922679 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.036327 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data-custom\") pod \"919e76fa-7915-4a0f-8482-b827879dcacb\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.036447 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data\") pod \"919e76fa-7915-4a0f-8482-b827879dcacb\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.036478 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-combined-ca-bundle\") pod \"919e76fa-7915-4a0f-8482-b827879dcacb\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.036539 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919e76fa-7915-4a0f-8482-b827879dcacb-logs\") pod \"919e76fa-7915-4a0f-8482-b827879dcacb\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.036577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf66w\" (UniqueName: \"kubernetes.io/projected/919e76fa-7915-4a0f-8482-b827879dcacb-kube-api-access-bf66w\") pod \"919e76fa-7915-4a0f-8482-b827879dcacb\" (UID: \"919e76fa-7915-4a0f-8482-b827879dcacb\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.040044 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/919e76fa-7915-4a0f-8482-b827879dcacb-logs" (OuterVolumeSpecName: "logs") pod "919e76fa-7915-4a0f-8482-b827879dcacb" (UID: "919e76fa-7915-4a0f-8482-b827879dcacb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.047993 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.048042 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.048083 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.048678 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.048731 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" gracePeriod=600 Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.053577 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "919e76fa-7915-4a0f-8482-b827879dcacb" (UID: "919e76fa-7915-4a0f-8482-b827879dcacb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.055265 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/919e76fa-7915-4a0f-8482-b827879dcacb-kube-api-access-bf66w" (OuterVolumeSpecName: "kube-api-access-bf66w") pod "919e76fa-7915-4a0f-8482-b827879dcacb" (UID: "919e76fa-7915-4a0f-8482-b827879dcacb"). InnerVolumeSpecName "kube-api-access-bf66w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.074773 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "919e76fa-7915-4a0f-8482-b827879dcacb" (UID: "919e76fa-7915-4a0f-8482-b827879dcacb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.092622 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data" (OuterVolumeSpecName: "config-data") pod "919e76fa-7915-4a0f-8482-b827879dcacb" (UID: "919e76fa-7915-4a0f-8482-b827879dcacb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.139036 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.139355 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.139367 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919e76fa-7915-4a0f-8482-b827879dcacb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.139377 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/919e76fa-7915-4a0f-8482-b827879dcacb-logs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.139387 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf66w\" (UniqueName: \"kubernetes.io/projected/919e76fa-7915-4a0f-8482-b827879dcacb-kube-api-access-bf66w\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: E1205 11:34:14.176792 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.205939 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" exitCode=0 Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.206005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9"} Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.206159 4809 scope.go:117] "RemoveContainer" containerID="09f7d635e94f05ae8e8f470c73493c96f0ec6c0a4510de5411cdb6df882057b9" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.207409 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:34:14 crc kubenswrapper[4809]: E1205 11:34:14.208438 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.215821 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.215919 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69c49cc5f8-w57p4" event={"ID":"919e76fa-7915-4a0f-8482-b827879dcacb","Type":"ContainerDied","Data":"96e87795672f5b2e4cfc92bc314b4cd446ceb9e5767b729b4440e87f4c902eeb"} Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.235461 4809 generic.go:334] "Generic (PLEG): container finished" podID="cd1f3174-82d6-4264-8e70-01f2ea709403" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" exitCode=0 Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.235858 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.237247 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cd1f3174-82d6-4264-8e70-01f2ea709403","Type":"ContainerDied","Data":"236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6"} Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.258504 4809 scope.go:117] "RemoveContainer" containerID="bd369f5a80a27b3f543389e7e7b300b6bde366ca4736f6ed6dc27ce406afa252" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.287715 4809 scope.go:117] "RemoveContainer" containerID="c9e36f881b5958ccc505a0d69b7688ee45bfcd1841d3987cabc75687e7c369c0" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.290086 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-69c49cc5f8-w57p4"] Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.298952 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.301857 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-69c49cc5f8-w57p4"] Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.309740 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.314442 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.427613 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6cffd45dc7-dpz7j" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9696/\": dial tcp 10.217.0.154:9696: connect: connection refused" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.443043 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-combined-ca-bundle\") pod \"cd1f3174-82d6-4264-8e70-01f2ea709403\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.443082 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2wql\" (UniqueName: \"kubernetes.io/projected/cd1f3174-82d6-4264-8e70-01f2ea709403-kube-api-access-f2wql\") pod \"cd1f3174-82d6-4264-8e70-01f2ea709403\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.443148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-config-data\") pod \"cd1f3174-82d6-4264-8e70-01f2ea709403\" (UID: \"cd1f3174-82d6-4264-8e70-01f2ea709403\") " Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.446515 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1f3174-82d6-4264-8e70-01f2ea709403-kube-api-access-f2wql" (OuterVolumeSpecName: "kube-api-access-f2wql") pod "cd1f3174-82d6-4264-8e70-01f2ea709403" (UID: "cd1f3174-82d6-4264-8e70-01f2ea709403"). InnerVolumeSpecName "kube-api-access-f2wql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.462246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd1f3174-82d6-4264-8e70-01f2ea709403" (UID: "cd1f3174-82d6-4264-8e70-01f2ea709403"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.466488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-config-data" (OuterVolumeSpecName: "config-data") pod "cd1f3174-82d6-4264-8e70-01f2ea709403" (UID: "cd1f3174-82d6-4264-8e70-01f2ea709403"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.545094 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.545128 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1f3174-82d6-4264-8e70-01f2ea709403-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.545140 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2wql\" (UniqueName: \"kubernetes.io/projected/cd1f3174-82d6-4264-8e70-01f2ea709403-kube-api-access-f2wql\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.881385 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" path="/var/lib/kubelet/pods/919e76fa-7915-4a0f-8482-b827879dcacb/volumes" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.882100 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc828bad-c058-4a1c-aa44-18d607759d6b" path="/var/lib/kubelet/pods/cc828bad-c058-4a1c-aa44-18d607759d6b/volumes" Dec 05 11:34:14 crc kubenswrapper[4809]: I1205 11:34:14.882888 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" path="/var/lib/kubelet/pods/cfe522ab-449f-433e-a9bf-b33ed3b42595/volumes" Dec 05 11:34:15 crc kubenswrapper[4809]: I1205 11:34:15.252925 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cd1f3174-82d6-4264-8e70-01f2ea709403","Type":"ContainerDied","Data":"24a44518182efb547a6928bd33cd3f773d93467bbce99f6c8531966dee797c46"} Dec 05 11:34:15 crc kubenswrapper[4809]: I1205 11:34:15.253020 4809 scope.go:117] "RemoveContainer" containerID="236e96e34cc6411d49456093fdf294f9ce1323625a5839c8b851ccb41cc0e1d6" Dec 05 11:34:15 crc kubenswrapper[4809]: I1205 11:34:15.253224 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 11:34:15 crc kubenswrapper[4809]: I1205 11:34:15.279114 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:34:15 crc kubenswrapper[4809]: I1205 11:34:15.284205 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:34:15 crc kubenswrapper[4809]: E1205 11:34:15.666857 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:15 crc kubenswrapper[4809]: E1205 11:34:15.666972 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts podName:77f68706-76e6-4ee3-b27d-bb092ab68a30 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:23.666947185 +0000 UTC m=+1559.057923783 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts") pod "neutronee61-account-delete-dfhwt" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30") : configmap "openstack-scripts" not found Dec 05 11:34:15 crc kubenswrapper[4809]: E1205 11:34:15.667033 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:15 crc kubenswrapper[4809]: E1205 11:34:15.667106 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts podName:102c9fa9-707d-4918-9a5d-55d8d4c47c12 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:23.667086429 +0000 UTC m=+1559.058062997 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts") pod "cinder18b7-account-delete-jcsnx" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12") : configmap "openstack-scripts" not found Dec 05 11:34:16 crc kubenswrapper[4809]: E1205 11:34:16.684956 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:16 crc kubenswrapper[4809]: E1205 11:34:16.685480 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts podName:dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c nodeName:}" failed. No retries permitted until 2025-12-05 11:34:24.685448103 +0000 UTC m=+1560.076424871 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts") pod "novacell00f47-account-delete-qcdx8" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c") : configmap "openstack-scripts" not found Dec 05 11:34:16 crc kubenswrapper[4809]: E1205 11:34:16.786397 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:16 crc kubenswrapper[4809]: E1205 11:34:16.786489 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts podName:4ae41780-4bf1-4a48-bf11-74dadb59e387 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:24.78647171 +0000 UTC m=+1560.177448278 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts") pod "novaapi293e-account-delete-tb822" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387") : configmap "openstack-scripts" not found Dec 05 11:34:16 crc kubenswrapper[4809]: I1205 11:34:16.888594 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd1f3174-82d6-4264-8e70-01f2ea709403" path="/var/lib/kubelet/pods/cd1f3174-82d6-4264-8e70-01f2ea709403/volumes" Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.217811 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.218173 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.218518 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.218567 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.219721 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.221102 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.222204 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:17 crc kubenswrapper[4809]: E1205 11:34:17.222254 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.218792 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.220073 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.221596 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.221952 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.222139 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.224318 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.225869 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:22 crc kubenswrapper[4809]: E1205 11:34:22.225905 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.877013 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.995090 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-combined-ca-bundle\") pod \"b85b1950-6c23-424d-a724-a2a8cad9a05a\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.995145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz58g\" (UniqueName: \"kubernetes.io/projected/b85b1950-6c23-424d-a724-a2a8cad9a05a-kube-api-access-lz58g\") pod \"b85b1950-6c23-424d-a724-a2a8cad9a05a\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.995203 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-httpd-config\") pod \"b85b1950-6c23-424d-a724-a2a8cad9a05a\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.995233 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-public-tls-certs\") pod \"b85b1950-6c23-424d-a724-a2a8cad9a05a\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.995262 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-config\") pod \"b85b1950-6c23-424d-a724-a2a8cad9a05a\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.995309 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-internal-tls-certs\") pod \"b85b1950-6c23-424d-a724-a2a8cad9a05a\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " Dec 05 11:34:22 crc kubenswrapper[4809]: I1205 11:34:22.995360 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-ovndb-tls-certs\") pod \"b85b1950-6c23-424d-a724-a2a8cad9a05a\" (UID: \"b85b1950-6c23-424d-a724-a2a8cad9a05a\") " Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.001805 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85b1950-6c23-424d-a724-a2a8cad9a05a-kube-api-access-lz58g" (OuterVolumeSpecName: "kube-api-access-lz58g") pod "b85b1950-6c23-424d-a724-a2a8cad9a05a" (UID: "b85b1950-6c23-424d-a724-a2a8cad9a05a"). InnerVolumeSpecName "kube-api-access-lz58g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.002047 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b85b1950-6c23-424d-a724-a2a8cad9a05a" (UID: "b85b1950-6c23-424d-a724-a2a8cad9a05a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.044953 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b85b1950-6c23-424d-a724-a2a8cad9a05a" (UID: "b85b1950-6c23-424d-a724-a2a8cad9a05a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.049369 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-config" (OuterVolumeSpecName: "config") pod "b85b1950-6c23-424d-a724-a2a8cad9a05a" (UID: "b85b1950-6c23-424d-a724-a2a8cad9a05a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.059120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b85b1950-6c23-424d-a724-a2a8cad9a05a" (UID: "b85b1950-6c23-424d-a724-a2a8cad9a05a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.072063 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b85b1950-6c23-424d-a724-a2a8cad9a05a" (UID: "b85b1950-6c23-424d-a724-a2a8cad9a05a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.080877 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b85b1950-6c23-424d-a724-a2a8cad9a05a" (UID: "b85b1950-6c23-424d-a724-a2a8cad9a05a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.097719 4809 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.097824 4809 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.097913 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.097933 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz58g\" (UniqueName: \"kubernetes.io/projected/b85b1950-6c23-424d-a724-a2a8cad9a05a-kube-api-access-lz58g\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.098015 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.098034 4809 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.098110 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b85b1950-6c23-424d-a724-a2a8cad9a05a-config\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.359816 4809 generic.go:334] "Generic (PLEG): container finished" podID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerID="987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e" exitCode=0 Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.359869 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffd45dc7-dpz7j" event={"ID":"b85b1950-6c23-424d-a724-a2a8cad9a05a","Type":"ContainerDied","Data":"987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e"} Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.359900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cffd45dc7-dpz7j" event={"ID":"b85b1950-6c23-424d-a724-a2a8cad9a05a","Type":"ContainerDied","Data":"b714fdf18431b0dc2b38d662d0c2f76233ea078d9d327041645a4cc2e679b59a"} Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.359920 4809 scope.go:117] "RemoveContainer" containerID="0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.359926 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cffd45dc7-dpz7j" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.391896 4809 scope.go:117] "RemoveContainer" containerID="987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.409659 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cffd45dc7-dpz7j"] Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.418060 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6cffd45dc7-dpz7j"] Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.434189 4809 scope.go:117] "RemoveContainer" containerID="0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55" Dec 05 11:34:23 crc kubenswrapper[4809]: E1205 11:34:23.434858 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55\": container with ID starting with 0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55 not found: ID does not exist" containerID="0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.434903 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55"} err="failed to get container status \"0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55\": rpc error: code = NotFound desc = could not find container \"0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55\": container with ID starting with 0ecd9636ceba695ca472215aa8f93af7fc8f1f5d9dd922c5d95c07c79aa2db55 not found: ID does not exist" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.434933 4809 scope.go:117] "RemoveContainer" containerID="987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e" Dec 05 11:34:23 crc kubenswrapper[4809]: E1205 11:34:23.435558 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e\": container with ID starting with 987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e not found: ID does not exist" containerID="987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e" Dec 05 11:34:23 crc kubenswrapper[4809]: I1205 11:34:23.435649 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e"} err="failed to get container status \"987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e\": rpc error: code = NotFound desc = could not find container \"987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e\": container with ID starting with 987e666b757be214a3d2834ec0a65fe19a3a4a4f5f88eec6673edddf3a71701e not found: ID does not exist" Dec 05 11:34:23 crc kubenswrapper[4809]: E1205 11:34:23.715175 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:23 crc kubenswrapper[4809]: E1205 11:34:23.715254 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts podName:102c9fa9-707d-4918-9a5d-55d8d4c47c12 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:39.715233003 +0000 UTC m=+1575.106209571 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts") pod "cinder18b7-account-delete-jcsnx" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12") : configmap "openstack-scripts" not found Dec 05 11:34:23 crc kubenswrapper[4809]: E1205 11:34:23.715291 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:23 crc kubenswrapper[4809]: E1205 11:34:23.715421 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts podName:77f68706-76e6-4ee3-b27d-bb092ab68a30 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:39.715393647 +0000 UTC m=+1575.106370245 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts") pod "neutronee61-account-delete-dfhwt" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30") : configmap "openstack-scripts" not found Dec 05 11:34:24 crc kubenswrapper[4809]: E1205 11:34:24.732055 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:24 crc kubenswrapper[4809]: E1205 11:34:24.732431 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts podName:dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c nodeName:}" failed. No retries permitted until 2025-12-05 11:34:40.732415795 +0000 UTC m=+1576.123392353 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts") pod "novacell00f47-account-delete-qcdx8" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c") : configmap "openstack-scripts" not found Dec 05 11:34:24 crc kubenswrapper[4809]: E1205 11:34:24.833437 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:24 crc kubenswrapper[4809]: E1205 11:34:24.833513 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts podName:4ae41780-4bf1-4a48-bf11-74dadb59e387 nodeName:}" failed. No retries permitted until 2025-12-05 11:34:40.833497643 +0000 UTC m=+1576.224474201 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts") pod "novaapi293e-account-delete-tb822" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387") : configmap "openstack-scripts" not found Dec 05 11:34:24 crc kubenswrapper[4809]: I1205 11:34:24.879296 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:34:24 crc kubenswrapper[4809]: E1205 11:34:24.879586 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:34:24 crc kubenswrapper[4809]: I1205 11:34:24.884495 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" path="/var/lib/kubelet/pods/b85b1950-6c23-424d-a724-a2a8cad9a05a/volumes" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.633248 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7xd65"] Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634117 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="sg-core" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634140 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="sg-core" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634160 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1f3174-82d6-4264-8e70-01f2ea709403" containerName="nova-cell1-conductor-conductor" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634194 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1f3174-82d6-4264-8e70-01f2ea709403" containerName="nova-cell1-conductor-conductor" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634232 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634269 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634330 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634359 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634395 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerName="mysql-bootstrap" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634429 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerName="mysql-bootstrap" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634495 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e68543-4956-4fc0-86aa-9ae8484b07c3" containerName="kube-state-metrics" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634527 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e68543-4956-4fc0-86aa-9ae8484b07c3" containerName="kube-state-metrics" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634563 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447a5847-564c-4b01-a74d-503249e0dd95" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634593 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="447a5847-564c-4b01-a74d-503249e0dd95" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634765 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerName="setup-container" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634813 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerName="setup-container" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634856 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634887 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.634944 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.634973 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635011 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerName="rabbitmq" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635044 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerName="rabbitmq" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635104 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="openstack-network-exporter" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635136 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="openstack-network-exporter" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635177 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="proxy-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635207 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="proxy-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635255 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635280 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635483 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-notification-agent" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635515 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-notification-agent" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635549 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635579 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635622 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635693 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635742 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635777 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635811 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="ovn-northd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635840 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="ovn-northd" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635882 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b3a18d-30f0-496f-9990-dd453ea8dbb7" containerName="nova-cell0-conductor-conductor" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635911 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b3a18d-30f0-496f-9990-dd453ea8dbb7" containerName="nova-cell0-conductor-conductor" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.635969 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.635987 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636011 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-central-agent" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636026 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-central-agent" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636046 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f96a466a-b1fb-4260-b831-e4cd814a767e" containerName="keystone-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636061 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f96a466a-b1fb-4260-b831-e4cd814a767e" containerName="keystone-api" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636076 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636089 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636106 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-metadata" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636118 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-metadata" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636139 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e409c78c-5de9-48ce-9c8b-afc67c6f8f48" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636152 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e409c78c-5de9-48ce-9c8b-afc67c6f8f48" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636185 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerName="galera" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636199 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerName="galera" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636222 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6712dcd8-a4a4-4f57-adee-6b218823ba77" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636238 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6712dcd8-a4a4-4f57-adee-6b218823ba77" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636256 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636274 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636299 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636311 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636334 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636346 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636363 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636375 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-api" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636393 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerName="mysql-bootstrap" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636406 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerName="mysql-bootstrap" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636424 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc828bad-c058-4a1c-aa44-18d607759d6b" containerName="nova-scheduler-scheduler" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636437 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc828bad-c058-4a1c-aa44-18d607759d6b" containerName="nova-scheduler-scheduler" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636461 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636473 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636491 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerName="galera" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636504 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerName="galera" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636521 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636534 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-api" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636556 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerName="rabbitmq" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636568 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerName="rabbitmq" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636585 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerName="setup-container" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636597 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerName="setup-container" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636617 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636654 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636670 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636683 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-log" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636708 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636722 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-api" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636741 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-server" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636756 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-server" Dec 05 11:34:26 crc kubenswrapper[4809]: E1205 11:34:26.636775 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c32262-ad33-4c11-9536-2b094ef77d55" containerName="memcached" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.636789 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c32262-ad33-4c11-9536-2b094ef77d55" containerName="memcached" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637072 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e409c78c-5de9-48ce-9c8b-afc67c6f8f48" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637101 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637120 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637143 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637160 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637181 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-metadata" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637199 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637222 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe522ab-449f-433e-a9bf-b33ed3b42595" containerName="rabbitmq" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637237 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="447a5847-564c-4b01-a74d-503249e0dd95" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637260 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637277 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-notification-agent" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637300 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b760a19-f9d4-4ea7-a6c7-6610aa0334a9" containerName="galera" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637322 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1251531-cb99-418e-aa38-65a793e5a9d0" containerName="rabbitmq" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637335 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="sg-core" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637353 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="openstack-network-exporter" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637377 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637403 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1f3174-82d6-4264-8e70-01f2ea709403" containerName="nova-cell1-conductor-conductor" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637425 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85b1950-6c23-424d-a724-a2a8cad9a05a" containerName="neutron-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637449 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="ceilometer-central-agent" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637463 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8226c08-75fc-4250-bc4d-e0dbbc818730" containerName="ovn-controller" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637485 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c32262-ad33-4c11-9536-2b094ef77d55" containerName="memcached" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637501 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="232c2be5-1e83-4fc2-b2c7-9a668d572d19" containerName="placement-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637522 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e68543-4956-4fc0-86aa-9ae8484b07c3" containerName="kube-state-metrics" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637545 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637566 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdb0190-0a3d-482e-acbc-5d40e7999345" containerName="proxy-server" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637583 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637603 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="755354bd-175b-46e6-a9bf-93d8b9eab885" containerName="barbican-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637623 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8220cf04-d91a-45d8-ad53-51d0c6d21fe5" containerName="nova-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637669 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf5e0d0-86bc-4728-8ad4-678b4b28e190" containerName="glance-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637686 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aabcdcf-4f51-4398-ae40-33b6d57aa514" containerName="ovn-northd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637702 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="846a6cff-41b5-4b49-8569-8c757669fe7f" containerName="cinder-api-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637725 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="871c924c-f749-40cd-81b2-64af559e6b20" containerName="nova-metadata-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637744 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc828bad-c058-4a1c-aa44-18d607759d6b" containerName="nova-scheduler-scheduler" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637759 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bb75c0-16a0-43d2-a344-aff2cd60b3c2" containerName="proxy-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637778 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2b5c2a4-1802-4308-9db8-090ea111f7ba" containerName="galera" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637792 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="919e76fa-7915-4a0f-8482-b827879dcacb" containerName="barbican-keystone-listener-log" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637813 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8da287-1c2a-4d39-9d1b-dc5218b44907" containerName="glance-httpd" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637828 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b3a18d-30f0-496f-9990-dd453ea8dbb7" containerName="nova-cell0-conductor-conductor" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637872 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f96a466a-b1fb-4260-b831-e4cd814a767e" containerName="keystone-api" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.637890 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6712dcd8-a4a4-4f57-adee-6b218823ba77" containerName="mariadb-account-delete" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.639977 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.651238 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7xd65"] Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.764316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-catalog-content\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.764521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhkdz\" (UniqueName: \"kubernetes.io/projected/e9c99d17-0a97-4f67-8295-68683c42a2c0-kube-api-access-zhkdz\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.764581 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-utilities\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.865805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhkdz\" (UniqueName: \"kubernetes.io/projected/e9c99d17-0a97-4f67-8295-68683c42a2c0-kube-api-access-zhkdz\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.865866 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-utilities\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.865984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-catalog-content\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.866546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-catalog-content\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.866712 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-utilities\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.884649 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhkdz\" (UniqueName: \"kubernetes.io/projected/e9c99d17-0a97-4f67-8295-68683c42a2c0-kube-api-access-zhkdz\") pod \"community-operators-7xd65\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:26 crc kubenswrapper[4809]: I1205 11:34:26.970170 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.218616 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.219407 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.219735 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.219775 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.219915 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.221185 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.224283 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:27 crc kubenswrapper[4809]: E1205 11:34:27.224344 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:34:27 crc kubenswrapper[4809]: I1205 11:34:27.494548 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7xd65"] Dec 05 11:34:28 crc kubenswrapper[4809]: I1205 11:34:28.416348 4809 generic.go:334] "Generic (PLEG): container finished" podID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerID="d1813eb7ebd4c84ef860894629d14373a23fbe561be07a13b8ab97ce6489b5d6" exitCode=0 Dec 05 11:34:28 crc kubenswrapper[4809]: I1205 11:34:28.416434 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xd65" event={"ID":"e9c99d17-0a97-4f67-8295-68683c42a2c0","Type":"ContainerDied","Data":"d1813eb7ebd4c84ef860894629d14373a23fbe561be07a13b8ab97ce6489b5d6"} Dec 05 11:34:28 crc kubenswrapper[4809]: I1205 11:34:28.417713 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xd65" event={"ID":"e9c99d17-0a97-4f67-8295-68683c42a2c0","Type":"ContainerStarted","Data":"474f0054191332866a1d7a234ed4d1c66d8783de5768529081c011cc0fa47bfc"} Dec 05 11:34:29 crc kubenswrapper[4809]: I1205 11:34:29.429154 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xd65" event={"ID":"e9c99d17-0a97-4f67-8295-68683c42a2c0","Type":"ContainerStarted","Data":"5a8fb2dd7cf19a75cf006c056a0b6fde4bcfdab15d1b4290a94a75f3974790a8"} Dec 05 11:34:30 crc kubenswrapper[4809]: I1205 11:34:30.439792 4809 generic.go:334] "Generic (PLEG): container finished" podID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerID="5a8fb2dd7cf19a75cf006c056a0b6fde4bcfdab15d1b4290a94a75f3974790a8" exitCode=0 Dec 05 11:34:30 crc kubenswrapper[4809]: I1205 11:34:30.439831 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xd65" event={"ID":"e9c99d17-0a97-4f67-8295-68683c42a2c0","Type":"ContainerDied","Data":"5a8fb2dd7cf19a75cf006c056a0b6fde4bcfdab15d1b4290a94a75f3974790a8"} Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.217834 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.218625 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.219069 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.219135 4809 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.219823 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.222442 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.224400 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 05 11:34:32 crc kubenswrapper[4809]: E1205 11:34:32.224445 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-bz8bw" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:34:32 crc kubenswrapper[4809]: I1205 11:34:32.459586 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xd65" event={"ID":"e9c99d17-0a97-4f67-8295-68683c42a2c0","Type":"ContainerStarted","Data":"19fdcf63c438ec877f737feb8122044e26464a3058b8769da85a495e14527c25"} Dec 05 11:34:32 crc kubenswrapper[4809]: I1205 11:34:32.478718 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7xd65" podStartSLOduration=3.5932731049999997 podStartE2EDuration="6.47869347s" podCreationTimestamp="2025-12-05 11:34:26 +0000 UTC" firstStartedPulling="2025-12-05 11:34:28.418404964 +0000 UTC m=+1563.809381542" lastFinishedPulling="2025-12-05 11:34:31.303825349 +0000 UTC m=+1566.694801907" observedRunningTime="2025-12-05 11:34:32.476444972 +0000 UTC m=+1567.867421550" watchObservedRunningTime="2025-12-05 11:34:32.47869347 +0000 UTC m=+1567.869670028" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.624111 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vpwkx"] Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.627096 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.632275 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vpwkx"] Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.710225 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b892319b-0c09-48a4-85af-dae860c7c31e-catalog-content\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.710300 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b892319b-0c09-48a4-85af-dae860c7c31e-utilities\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.710335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qbj4\" (UniqueName: \"kubernetes.io/projected/b892319b-0c09-48a4-85af-dae860c7c31e-kube-api-access-6qbj4\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.811240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b892319b-0c09-48a4-85af-dae860c7c31e-catalog-content\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.812116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b892319b-0c09-48a4-85af-dae860c7c31e-utilities\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.812040 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b892319b-0c09-48a4-85af-dae860c7c31e-catalog-content\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.812188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qbj4\" (UniqueName: \"kubernetes.io/projected/b892319b-0c09-48a4-85af-dae860c7c31e-kube-api-access-6qbj4\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.812415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b892319b-0c09-48a4-85af-dae860c7c31e-utilities\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.833424 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qbj4\" (UniqueName: \"kubernetes.io/projected/b892319b-0c09-48a4-85af-dae860c7c31e-kube-api-access-6qbj4\") pod \"redhat-marketplace-vpwkx\" (UID: \"b892319b-0c09-48a4-85af-dae860c7c31e\") " pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:34 crc kubenswrapper[4809]: I1205 11:34:34.967780 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.446473 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vpwkx"] Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.495203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vpwkx" event={"ID":"b892319b-0c09-48a4-85af-dae860c7c31e","Type":"ContainerStarted","Data":"ac187b3592114cf82307f32da46182e161d9ce54ea941d3f5fedb45eb763aa21"} Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.507371 4809 generic.go:334] "Generic (PLEG): container finished" podID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerID="0414b93f70cfeebd7e27ab579af01c51e47ad9419cbb11d16f41369b2fe5cc96" exitCode=137 Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.507460 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"0414b93f70cfeebd7e27ab579af01c51e47ad9419cbb11d16f41369b2fe5cc96"} Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.509742 4809 generic.go:334] "Generic (PLEG): container finished" podID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerID="ed7d88f2530268a3f4438c5034a010adcd8f5b9f881c741aaee213b005a6a326" exitCode=137 Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.509817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3d12a95a-49fa-4392-9a0d-9efd9aa28720","Type":"ContainerDied","Data":"ed7d88f2530268a3f4438c5034a010adcd8f5b9f881c741aaee213b005a6a326"} Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.511693 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bz8bw_11752f85-2262-477b-9920-bf5c13f45f20/ovs-vswitchd/0.log" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.512551 4809 generic.go:334] "Generic (PLEG): container finished" podID="11752f85-2262-477b-9920-bf5c13f45f20" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" exitCode=137 Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.512589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bz8bw" event={"ID":"11752f85-2262-477b-9920-bf5c13f45f20","Type":"ContainerDied","Data":"5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4"} Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.601526 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.623257 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data\") pod \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.623305 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d12a95a-49fa-4392-9a0d-9efd9aa28720-etc-machine-id\") pod \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.623365 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbl6g\" (UniqueName: \"kubernetes.io/projected/3d12a95a-49fa-4392-9a0d-9efd9aa28720-kube-api-access-hbl6g\") pod \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.623394 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-combined-ca-bundle\") pod \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.623478 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data-custom\") pod \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.623513 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-scripts\") pod \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\" (UID: \"3d12a95a-49fa-4392-9a0d-9efd9aa28720\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.626784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3d12a95a-49fa-4392-9a0d-9efd9aa28720-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3d12a95a-49fa-4392-9a0d-9efd9aa28720" (UID: "3d12a95a-49fa-4392-9a0d-9efd9aa28720"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.629910 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-scripts" (OuterVolumeSpecName: "scripts") pod "3d12a95a-49fa-4392-9a0d-9efd9aa28720" (UID: "3d12a95a-49fa-4392-9a0d-9efd9aa28720"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.630168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3d12a95a-49fa-4392-9a0d-9efd9aa28720" (UID: "3d12a95a-49fa-4392-9a0d-9efd9aa28720"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.630824 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d12a95a-49fa-4392-9a0d-9efd9aa28720-kube-api-access-hbl6g" (OuterVolumeSpecName: "kube-api-access-hbl6g") pod "3d12a95a-49fa-4392-9a0d-9efd9aa28720" (UID: "3d12a95a-49fa-4392-9a0d-9efd9aa28720"). InnerVolumeSpecName "kube-api-access-hbl6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.667494 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d12a95a-49fa-4392-9a0d-9efd9aa28720" (UID: "3d12a95a-49fa-4392-9a0d-9efd9aa28720"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.702756 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data" (OuterVolumeSpecName: "config-data") pod "3d12a95a-49fa-4392-9a0d-9efd9aa28720" (UID: "3d12a95a-49fa-4392-9a0d-9efd9aa28720"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.725649 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.725976 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d12a95a-49fa-4392-9a0d-9efd9aa28720-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.725993 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbl6g\" (UniqueName: \"kubernetes.io/projected/3d12a95a-49fa-4392-9a0d-9efd9aa28720-kube-api-access-hbl6g\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.726007 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.726020 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.726031 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d12a95a-49fa-4392-9a0d-9efd9aa28720-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.776732 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bz8bw_11752f85-2262-477b-9920-bf5c13f45f20/ovs-vswitchd/0.log" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.777801 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826239 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11752f85-2262-477b-9920-bf5c13f45f20-scripts\") pod \"11752f85-2262-477b-9920-bf5c13f45f20\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826295 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-run\") pod \"11752f85-2262-477b-9920-bf5c13f45f20\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826339 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-lib\") pod \"11752f85-2262-477b-9920-bf5c13f45f20\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826358 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp5tq\" (UniqueName: \"kubernetes.io/projected/11752f85-2262-477b-9920-bf5c13f45f20-kube-api-access-mp5tq\") pod \"11752f85-2262-477b-9920-bf5c13f45f20\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-log\") pod \"11752f85-2262-477b-9920-bf5c13f45f20\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826419 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-etc-ovs\") pod \"11752f85-2262-477b-9920-bf5c13f45f20\" (UID: \"11752f85-2262-477b-9920-bf5c13f45f20\") " Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826723 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "11752f85-2262-477b-9920-bf5c13f45f20" (UID: "11752f85-2262-477b-9920-bf5c13f45f20"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826765 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-run" (OuterVolumeSpecName: "var-run") pod "11752f85-2262-477b-9920-bf5c13f45f20" (UID: "11752f85-2262-477b-9920-bf5c13f45f20"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.826786 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-lib" (OuterVolumeSpecName: "var-lib") pod "11752f85-2262-477b-9920-bf5c13f45f20" (UID: "11752f85-2262-477b-9920-bf5c13f45f20"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.827268 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-log" (OuterVolumeSpecName: "var-log") pod "11752f85-2262-477b-9920-bf5c13f45f20" (UID: "11752f85-2262-477b-9920-bf5c13f45f20"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.827561 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11752f85-2262-477b-9920-bf5c13f45f20-scripts" (OuterVolumeSpecName: "scripts") pod "11752f85-2262-477b-9920-bf5c13f45f20" (UID: "11752f85-2262-477b-9920-bf5c13f45f20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.835305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11752f85-2262-477b-9920-bf5c13f45f20-kube-api-access-mp5tq" (OuterVolumeSpecName: "kube-api-access-mp5tq") pod "11752f85-2262-477b-9920-bf5c13f45f20" (UID: "11752f85-2262-477b-9920-bf5c13f45f20"). InnerVolumeSpecName "kube-api-access-mp5tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.927819 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11752f85-2262-477b-9920-bf5c13f45f20-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.927847 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.927856 4809 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-lib\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.927865 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp5tq\" (UniqueName: \"kubernetes.io/projected/11752f85-2262-477b-9920-bf5c13f45f20-kube-api-access-mp5tq\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.927873 4809 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-var-log\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:35 crc kubenswrapper[4809]: I1205 11:34:35.927882 4809 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/11752f85-2262-477b-9920-bf5c13f45f20-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.174658 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.333978 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") pod \"58501ab1-18d7-45d9-98cb-a74e62db36fc\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.334120 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"58501ab1-18d7-45d9-98cb-a74e62db36fc\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.334243 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-lock\") pod \"58501ab1-18d7-45d9-98cb-a74e62db36fc\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.334282 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrb7j\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-kube-api-access-mrb7j\") pod \"58501ab1-18d7-45d9-98cb-a74e62db36fc\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.334312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-cache\") pod \"58501ab1-18d7-45d9-98cb-a74e62db36fc\" (UID: \"58501ab1-18d7-45d9-98cb-a74e62db36fc\") " Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.334804 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-lock" (OuterVolumeSpecName: "lock") pod "58501ab1-18d7-45d9-98cb-a74e62db36fc" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.334834 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-cache" (OuterVolumeSpecName: "cache") pod "58501ab1-18d7-45d9-98cb-a74e62db36fc" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.337414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "58501ab1-18d7-45d9-98cb-a74e62db36fc" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.337550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-kube-api-access-mrb7j" (OuterVolumeSpecName: "kube-api-access-mrb7j") pod "58501ab1-18d7-45d9-98cb-a74e62db36fc" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc"). InnerVolumeSpecName "kube-api-access-mrb7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.337702 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "swift") pod "58501ab1-18d7-45d9-98cb-a74e62db36fc" (UID: "58501ab1-18d7-45d9-98cb-a74e62db36fc"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.436933 4809 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-lock\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.436996 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrb7j\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-kube-api-access-mrb7j\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.437022 4809 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/58501ab1-18d7-45d9-98cb-a74e62db36fc-cache\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.437044 4809 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/58501ab1-18d7-45d9-98cb-a74e62db36fc-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.437098 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.469604 4809 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.526408 4809 generic.go:334] "Generic (PLEG): container finished" podID="b892319b-0c09-48a4-85af-dae860c7c31e" containerID="d040bef640c71d50aae348bf50de89806ac896c2867edfce937853571aab6254" exitCode=0 Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.526517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vpwkx" event={"ID":"b892319b-0c09-48a4-85af-dae860c7c31e","Type":"ContainerDied","Data":"d040bef640c71d50aae348bf50de89806ac896c2867edfce937853571aab6254"} Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.539141 4809 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.541292 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"58501ab1-18d7-45d9-98cb-a74e62db36fc","Type":"ContainerDied","Data":"ebdbfd0085c5a8b87e7a702a3c0c3d81bf14bb2eec0523014bd68a71b4c5e402"} Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.541356 4809 scope.go:117] "RemoveContainer" containerID="0414b93f70cfeebd7e27ab579af01c51e47ad9419cbb11d16f41369b2fe5cc96" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.541463 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.544537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3d12a95a-49fa-4392-9a0d-9efd9aa28720","Type":"ContainerDied","Data":"d7a7389b0b4ebe9335b1d9fe477e4e17ba0dda482d71572edcbd983856081ad2"} Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.544939 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.547505 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bz8bw_11752f85-2262-477b-9920-bf5c13f45f20/ovs-vswitchd/0.log" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.549756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bz8bw" event={"ID":"11752f85-2262-477b-9920-bf5c13f45f20","Type":"ContainerDied","Data":"9adf06c9ae9df7ca21d5ede2cfc3fc966b7d0e941944b2e08a6149a9954c264f"} Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.549949 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bz8bw" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.584919 4809 scope.go:117] "RemoveContainer" containerID="e13456d6c296dda4598f34ee0b8c0e4703c954f9a8553fd24476fb75c023f44c" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.617936 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.620124 4809 scope.go:117] "RemoveContainer" containerID="9d5533e9d925f25fad5f31fbd6f5b9fe99bee0d29f635ea849b663a235a46895" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.631484 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.654510 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.667626 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.669040 4809 scope.go:117] "RemoveContainer" containerID="640615ea9e04efceac48bfd84c7eaa5885da503591e80a899131fe2037fed294" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.672767 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-bz8bw"] Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.677941 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-bz8bw"] Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.686325 4809 scope.go:117] "RemoveContainer" containerID="22c2f8cd7c22d4a0109b044a886f55e278f8c205daf14ca5e1b47bd41b23426a" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.702507 4809 scope.go:117] "RemoveContainer" containerID="6bb7e4f8fe910073653aa1a9f3a5a433f92b9f57a52d341885532226825ec2ca" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.760992 4809 scope.go:117] "RemoveContainer" containerID="fc97466f1d8534b47cb24e449245a63275f1660daf521b08f83aaabeed242660" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.784077 4809 scope.go:117] "RemoveContainer" containerID="213d6625154c146c9050604a6a30c02a271e04a3445369142a80d5fbd10042c8" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.811087 4809 scope.go:117] "RemoveContainer" containerID="f644863166060fc77243c2059ca8902c14f11a6c9b85f76eb1d36cd09decb692" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.833986 4809 scope.go:117] "RemoveContainer" containerID="d80d9b0fd1b858fe84af319a01ca8ae68a9151b6c070e69dff8050d858ebc3be" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.862111 4809 scope.go:117] "RemoveContainer" containerID="58ee5b8b26614f934ab1ca4e41f1129f8f4b37fb56c414f5eef99f29744b4275" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.883357 4809 scope.go:117] "RemoveContainer" containerID="c952b386502163d911d4615800b364440abeafbe8af4d4762f426ad2d8920b4c" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.885204 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11752f85-2262-477b-9920-bf5c13f45f20" path="/var/lib/kubelet/pods/11752f85-2262-477b-9920-bf5c13f45f20/volumes" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.886571 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" path="/var/lib/kubelet/pods/3d12a95a-49fa-4392-9a0d-9efd9aa28720/volumes" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.887523 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" path="/var/lib/kubelet/pods/58501ab1-18d7-45d9-98cb-a74e62db36fc/volumes" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.903397 4809 scope.go:117] "RemoveContainer" containerID="0bb5aa4d3d56722e820bbab3a1408f0e2a8ad15ee9a7ade8db2984056113caab" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.925284 4809 scope.go:117] "RemoveContainer" containerID="2af085bac46d211c6f1357baf53c68b8c882c87ee5a214e91c8e24cfb34aaa57" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.941511 4809 scope.go:117] "RemoveContainer" containerID="5cdc142e3c8a66238160b2d52e5885ba4ddb6262d5ad5802d543fa708c536590" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.961534 4809 scope.go:117] "RemoveContainer" containerID="76602c05e47629f8dd3a939796b77e7372f979cf7d26325edfe41b5b970ca71c" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.971237 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.971378 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.979588 4809 scope.go:117] "RemoveContainer" containerID="ed7d88f2530268a3f4438c5034a010adcd8f5b9f881c741aaee213b005a6a326" Dec 05 11:34:36 crc kubenswrapper[4809]: I1205 11:34:36.997892 4809 scope.go:117] "RemoveContainer" containerID="5573c486e96396a45ca3302df8a9fd4a3516c6f86f4e5b8e79e4b5e2b5e1a4a4" Dec 05 11:34:37 crc kubenswrapper[4809]: I1205 11:34:37.015768 4809 scope.go:117] "RemoveContainer" containerID="911bbd5ec6762cccda501aa10d40c22cff3d9ba0161b985def969845da0f6f62" Dec 05 11:34:37 crc kubenswrapper[4809]: I1205 11:34:37.030365 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:37 crc kubenswrapper[4809]: I1205 11:34:37.034257 4809 scope.go:117] "RemoveContainer" containerID="a62571ec4d894419275eeeda6a564866e0faca31c38feba3449fddfeb7025a82" Dec 05 11:34:37 crc kubenswrapper[4809]: I1205 11:34:37.638806 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:37 crc kubenswrapper[4809]: I1205 11:34:37.872553 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:34:37 crc kubenswrapper[4809]: E1205 11:34:37.872890 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:34:39 crc kubenswrapper[4809]: I1205 11:34:39.181817 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7xd65"] Dec 05 11:34:39 crc kubenswrapper[4809]: E1205 11:34:39.796446 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:39 crc kubenswrapper[4809]: E1205 11:34:39.796656 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:39 crc kubenswrapper[4809]: E1205 11:34:39.797145 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts podName:102c9fa9-707d-4918-9a5d-55d8d4c47c12 nodeName:}" failed. No retries permitted until 2025-12-05 11:35:11.797117057 +0000 UTC m=+1607.188093645 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts") pod "cinder18b7-account-delete-jcsnx" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12") : configmap "openstack-scripts" not found Dec 05 11:34:39 crc kubenswrapper[4809]: E1205 11:34:39.799693 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts podName:77f68706-76e6-4ee3-b27d-bb092ab68a30 nodeName:}" failed. No retries permitted until 2025-12-05 11:35:11.799671083 +0000 UTC m=+1607.190647681 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts") pod "neutronee61-account-delete-dfhwt" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30") : configmap "openstack-scripts" not found Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.424530 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:40 crc kubenswrapper[4809]: E1205 11:34:40.501583 4809 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/b00850ad132ba7180e026ca582b2dd27c28d16e247590a02d0ac2c7d2ad6df03/diff" to get inode usage: stat /var/lib/containers/storage/overlay/b00850ad132ba7180e026ca582b2dd27c28d16e247590a02d0ac2c7d2ad6df03/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_swift-storage-0_58501ab1-18d7-45d9-98cb-a74e62db36fc/swift-recon-cron/0.log" to get inode usage: stat /var/log/pods/openstack_swift-storage-0_58501ab1-18d7-45d9-98cb-a74e62db36fc/swift-recon-cron/0.log: no such file or directory Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.516476 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.596433 4809 generic.go:334] "Generic (PLEG): container finished" podID="b892319b-0c09-48a4-85af-dae860c7c31e" containerID="ea5c84649752374029d1fcbc17dd6ebc39681aeb12a8ba945059af1e1f1bf0c6" exitCode=0 Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.596537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vpwkx" event={"ID":"b892319b-0c09-48a4-85af-dae860c7c31e","Type":"ContainerDied","Data":"ea5c84649752374029d1fcbc17dd6ebc39681aeb12a8ba945059af1e1f1bf0c6"} Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.597699 4809 generic.go:334] "Generic (PLEG): container finished" podID="102c9fa9-707d-4918-9a5d-55d8d4c47c12" containerID="3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73" exitCode=137 Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.597753 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder18b7-account-delete-jcsnx" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.597758 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder18b7-account-delete-jcsnx" event={"ID":"102c9fa9-707d-4918-9a5d-55d8d4c47c12","Type":"ContainerDied","Data":"3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73"} Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.597865 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder18b7-account-delete-jcsnx" event={"ID":"102c9fa9-707d-4918-9a5d-55d8d4c47c12","Type":"ContainerDied","Data":"c431b321f3820e7c612b4f190e35b1e4791949bc83b859b29302c641fa6b67aa"} Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.597885 4809 scope.go:117] "RemoveContainer" containerID="3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.607023 4809 generic.go:334] "Generic (PLEG): container finished" podID="77f68706-76e6-4ee3-b27d-bb092ab68a30" containerID="4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8" exitCode=137 Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.607110 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronee61-account-delete-dfhwt" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.607107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronee61-account-delete-dfhwt" event={"ID":"77f68706-76e6-4ee3-b27d-bb092ab68a30","Type":"ContainerDied","Data":"4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8"} Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.607170 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronee61-account-delete-dfhwt" event={"ID":"77f68706-76e6-4ee3-b27d-bb092ab68a30","Type":"ContainerDied","Data":"7e890444d0b95881a1015bb83540266f0364b4b9917cc820522d9cbe6336eece"} Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.607236 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7xd65" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="registry-server" containerID="cri-o://19fdcf63c438ec877f737feb8122044e26464a3058b8769da85a495e14527c25" gracePeriod=2 Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.607272 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk9s6\" (UniqueName: \"kubernetes.io/projected/102c9fa9-707d-4918-9a5d-55d8d4c47c12-kube-api-access-zk9s6\") pod \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.607609 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts\") pod \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\" (UID: \"102c9fa9-707d-4918-9a5d-55d8d4c47c12\") " Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.608365 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "102c9fa9-707d-4918-9a5d-55d8d4c47c12" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.621028 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/102c9fa9-707d-4918-9a5d-55d8d4c47c12-kube-api-access-zk9s6" (OuterVolumeSpecName: "kube-api-access-zk9s6") pod "102c9fa9-707d-4918-9a5d-55d8d4c47c12" (UID: "102c9fa9-707d-4918-9a5d-55d8d4c47c12"). InnerVolumeSpecName "kube-api-access-zk9s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.622482 4809 scope.go:117] "RemoveContainer" containerID="3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73" Dec 05 11:34:40 crc kubenswrapper[4809]: E1205 11:34:40.623476 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73\": container with ID starting with 3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73 not found: ID does not exist" containerID="3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.623520 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73"} err="failed to get container status \"3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73\": rpc error: code = NotFound desc = could not find container \"3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73\": container with ID starting with 3cabb07a2de129eca0e4f42b9a477767096072fb20c712d4ab80c6aa142b1f73 not found: ID does not exist" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.623547 4809 scope.go:117] "RemoveContainer" containerID="4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.647881 4809 scope.go:117] "RemoveContainer" containerID="4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8" Dec 05 11:34:40 crc kubenswrapper[4809]: E1205 11:34:40.648275 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8\": container with ID starting with 4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8 not found: ID does not exist" containerID="4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.648321 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8"} err="failed to get container status \"4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8\": rpc error: code = NotFound desc = could not find container \"4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8\": container with ID starting with 4bd8379729c008a2b93c470c66c9c2a127dc42264037f07e36a4881c254192e8 not found: ID does not exist" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.709379 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts\") pod \"77f68706-76e6-4ee3-b27d-bb092ab68a30\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.709459 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hn6s\" (UniqueName: \"kubernetes.io/projected/77f68706-76e6-4ee3-b27d-bb092ab68a30-kube-api-access-5hn6s\") pod \"77f68706-76e6-4ee3-b27d-bb092ab68a30\" (UID: \"77f68706-76e6-4ee3-b27d-bb092ab68a30\") " Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.709808 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/102c9fa9-707d-4918-9a5d-55d8d4c47c12-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.709821 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk9s6\" (UniqueName: \"kubernetes.io/projected/102c9fa9-707d-4918-9a5d-55d8d4c47c12-kube-api-access-zk9s6\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.709874 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77f68706-76e6-4ee3-b27d-bb092ab68a30" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.711941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f68706-76e6-4ee3-b27d-bb092ab68a30-kube-api-access-5hn6s" (OuterVolumeSpecName: "kube-api-access-5hn6s") pod "77f68706-76e6-4ee3-b27d-bb092ab68a30" (UID: "77f68706-76e6-4ee3-b27d-bb092ab68a30"). InnerVolumeSpecName "kube-api-access-5hn6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.811132 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77f68706-76e6-4ee3-b27d-bb092ab68a30-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.811172 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hn6s\" (UniqueName: \"kubernetes.io/projected/77f68706-76e6-4ee3-b27d-bb092ab68a30-kube-api-access-5hn6s\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:40 crc kubenswrapper[4809]: E1205 11:34:40.811245 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:40 crc kubenswrapper[4809]: E1205 11:34:40.811311 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts podName:dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c nodeName:}" failed. No retries permitted until 2025-12-05 11:35:12.811290933 +0000 UTC m=+1608.202267501 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts") pod "novacell00f47-account-delete-qcdx8" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c") : configmap "openstack-scripts" not found Dec 05 11:34:40 crc kubenswrapper[4809]: E1205 11:34:40.912711 4809 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 05 11:34:40 crc kubenswrapper[4809]: E1205 11:34:40.912784 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts podName:4ae41780-4bf1-4a48-bf11-74dadb59e387 nodeName:}" failed. No retries permitted until 2025-12-05 11:35:12.912762981 +0000 UTC m=+1608.303739679 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts") pod "novaapi293e-account-delete-tb822" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387") : configmap "openstack-scripts" not found Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.939586 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder18b7-account-delete-jcsnx"] Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.946808 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder18b7-account-delete-jcsnx"] Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.953650 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronee61-account-delete-dfhwt"] Dec 05 11:34:40 crc kubenswrapper[4809]: I1205 11:34:40.959616 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronee61-account-delete-dfhwt"] Dec 05 11:34:41 crc kubenswrapper[4809]: W1205 11:34:41.210412 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-conmon-d040bef640c71d50aae348bf50de89806ac896c2867edfce937853571aab6254.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-conmon-d040bef640c71d50aae348bf50de89806ac896c2867edfce937853571aab6254.scope: no such file or directory Dec 05 11:34:41 crc kubenswrapper[4809]: W1205 11:34:41.210562 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-d040bef640c71d50aae348bf50de89806ac896c2867edfce937853571aab6254.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-d040bef640c71d50aae348bf50de89806ac896c2867edfce937853571aab6254.scope: no such file or directory Dec 05 11:34:41 crc kubenswrapper[4809]: W1205 11:34:41.213434 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-conmon-ea5c84649752374029d1fcbc17dd6ebc39681aeb12a8ba945059af1e1f1bf0c6.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-conmon-ea5c84649752374029d1fcbc17dd6ebc39681aeb12a8ba945059af1e1f1bf0c6.scope: no such file or directory Dec 05 11:34:41 crc kubenswrapper[4809]: W1205 11:34:41.213561 4809 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-ea5c84649752374029d1fcbc17dd6ebc39681aeb12a8ba945059af1e1f1bf0c6.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb892319b_0c09_48a4_85af_dae860c7c31e.slice/crio-ea5c84649752374029d1fcbc17dd6ebc39681aeb12a8ba945059af1e1f1bf0c6.scope: no such file or directory Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.602986 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.647939 4809 generic.go:334] "Generic (PLEG): container finished" podID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerID="19fdcf63c438ec877f737feb8122044e26464a3058b8769da85a495e14527c25" exitCode=0 Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.647993 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xd65" event={"ID":"e9c99d17-0a97-4f67-8295-68683c42a2c0","Type":"ContainerDied","Data":"19fdcf63c438ec877f737feb8122044e26464a3058b8769da85a495e14527c25"} Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.650369 4809 generic.go:334] "Generic (PLEG): container finished" podID="4ae41780-4bf1-4a48-bf11-74dadb59e387" containerID="1d84c3afeb7263e6079e6e47565f86fd4f1c3311f83150774d2707c65bc27b18" exitCode=137 Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.650411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi293e-account-delete-tb822" event={"ID":"4ae41780-4bf1-4a48-bf11-74dadb59e387","Type":"ContainerDied","Data":"1d84c3afeb7263e6079e6e47565f86fd4f1c3311f83150774d2707c65bc27b18"} Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.655294 4809 generic.go:334] "Generic (PLEG): container finished" podID="dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" containerID="d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18" exitCode=137 Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.655381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell00f47-account-delete-qcdx8" event={"ID":"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c","Type":"ContainerDied","Data":"d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18"} Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.655377 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell00f47-account-delete-qcdx8" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.655408 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell00f47-account-delete-qcdx8" event={"ID":"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c","Type":"ContainerDied","Data":"f1c13975c52f4b444845b4d98d685f5d9795f21536d4f0da75d936d87b8024ca"} Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.655427 4809 scope.go:117] "RemoveContainer" containerID="d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.658517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vpwkx" event={"ID":"b892319b-0c09-48a4-85af-dae860c7c31e","Type":"ContainerStarted","Data":"e1232e49219becef533cc49cedf18f5b1207693dab2613c9e200fb49c35aebb6"} Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.692364 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vpwkx" podStartSLOduration=3.321163676 podStartE2EDuration="8.692333524s" podCreationTimestamp="2025-12-05 11:34:34 +0000 UTC" firstStartedPulling="2025-12-05 11:34:36.528845215 +0000 UTC m=+1571.919821823" lastFinishedPulling="2025-12-05 11:34:41.900015103 +0000 UTC m=+1577.290991671" observedRunningTime="2025-12-05 11:34:42.679690498 +0000 UTC m=+1578.070667066" watchObservedRunningTime="2025-12-05 11:34:42.692333524 +0000 UTC m=+1578.083310122" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.726594 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.733967 4809 scope.go:117] "RemoveContainer" containerID="d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18" Dec 05 11:34:42 crc kubenswrapper[4809]: E1205 11:34:42.735986 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18\": container with ID starting with d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18 not found: ID does not exist" containerID="d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.736019 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18"} err="failed to get container status \"d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18\": rpc error: code = NotFound desc = could not find container \"d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18\": container with ID starting with d7c1117da36aea8dfd2cf84554648a3fa34624980e33d076e124b8aa849a0c18 not found: ID does not exist" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.741584 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpt52\" (UniqueName: \"kubernetes.io/projected/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-kube-api-access-tpt52\") pod \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.741683 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts\") pod \"4ae41780-4bf1-4a48-bf11-74dadb59e387\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.741711 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts\") pod \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\" (UID: \"dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c\") " Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.741744 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxqvc\" (UniqueName: \"kubernetes.io/projected/4ae41780-4bf1-4a48-bf11-74dadb59e387-kube-api-access-rxqvc\") pod \"4ae41780-4bf1-4a48-bf11-74dadb59e387\" (UID: \"4ae41780-4bf1-4a48-bf11-74dadb59e387\") " Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.742849 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ae41780-4bf1-4a48-bf11-74dadb59e387" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.743024 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.748548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-kube-api-access-tpt52" (OuterVolumeSpecName: "kube-api-access-tpt52") pod "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" (UID: "dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c"). InnerVolumeSpecName "kube-api-access-tpt52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.748618 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ae41780-4bf1-4a48-bf11-74dadb59e387-kube-api-access-rxqvc" (OuterVolumeSpecName: "kube-api-access-rxqvc") pod "4ae41780-4bf1-4a48-bf11-74dadb59e387" (UID: "4ae41780-4bf1-4a48-bf11-74dadb59e387"). InnerVolumeSpecName "kube-api-access-rxqvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.831285 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.845276 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-utilities\") pod \"e9c99d17-0a97-4f67-8295-68683c42a2c0\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.845794 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpt52\" (UniqueName: \"kubernetes.io/projected/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-kube-api-access-tpt52\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.845812 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ae41780-4bf1-4a48-bf11-74dadb59e387-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.845824 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.845833 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxqvc\" (UniqueName: \"kubernetes.io/projected/4ae41780-4bf1-4a48-bf11-74dadb59e387-kube-api-access-rxqvc\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.846483 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-utilities" (OuterVolumeSpecName: "utilities") pod "e9c99d17-0a97-4f67-8295-68683c42a2c0" (UID: "e9c99d17-0a97-4f67-8295-68683c42a2c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.887446 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="102c9fa9-707d-4918-9a5d-55d8d4c47c12" path="/var/lib/kubelet/pods/102c9fa9-707d-4918-9a5d-55d8d4c47c12/volumes" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.888186 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f68706-76e6-4ee3-b27d-bb092ab68a30" path="/var/lib/kubelet/pods/77f68706-76e6-4ee3-b27d-bb092ab68a30/volumes" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.946607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhkdz\" (UniqueName: \"kubernetes.io/projected/e9c99d17-0a97-4f67-8295-68683c42a2c0-kube-api-access-zhkdz\") pod \"e9c99d17-0a97-4f67-8295-68683c42a2c0\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.946677 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-catalog-content\") pod \"e9c99d17-0a97-4f67-8295-68683c42a2c0\" (UID: \"e9c99d17-0a97-4f67-8295-68683c42a2c0\") " Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.947021 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.952835 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9c99d17-0a97-4f67-8295-68683c42a2c0-kube-api-access-zhkdz" (OuterVolumeSpecName: "kube-api-access-zhkdz") pod "e9c99d17-0a97-4f67-8295-68683c42a2c0" (UID: "e9c99d17-0a97-4f67-8295-68683c42a2c0"). InnerVolumeSpecName "kube-api-access-zhkdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.977389 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell00f47-account-delete-qcdx8"] Dec 05 11:34:42 crc kubenswrapper[4809]: I1205 11:34:42.983417 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell00f47-account-delete-qcdx8"] Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.002046 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9c99d17-0a97-4f67-8295-68683c42a2c0" (UID: "e9c99d17-0a97-4f67-8295-68683c42a2c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.047755 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhkdz\" (UniqueName: \"kubernetes.io/projected/e9c99d17-0a97-4f67-8295-68683c42a2c0-kube-api-access-zhkdz\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.047791 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c99d17-0a97-4f67-8295-68683c42a2c0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.670192 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xd65" event={"ID":"e9c99d17-0a97-4f67-8295-68683c42a2c0","Type":"ContainerDied","Data":"474f0054191332866a1d7a234ed4d1c66d8783de5768529081c011cc0fa47bfc"} Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.670571 4809 scope.go:117] "RemoveContainer" containerID="19fdcf63c438ec877f737feb8122044e26464a3058b8769da85a495e14527c25" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.670221 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xd65" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.676286 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi293e-account-delete-tb822" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.677118 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi293e-account-delete-tb822" event={"ID":"4ae41780-4bf1-4a48-bf11-74dadb59e387","Type":"ContainerDied","Data":"dfef12a118ca5df7c5dba0bd8130f9595102f2bfeef282b56dce6ba5a62b7226"} Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.697748 4809 scope.go:117] "RemoveContainer" containerID="5a8fb2dd7cf19a75cf006c056a0b6fde4bcfdab15d1b4290a94a75f3974790a8" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.709696 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi293e-account-delete-tb822"] Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.730231 4809 scope.go:117] "RemoveContainer" containerID="d1813eb7ebd4c84ef860894629d14373a23fbe561be07a13b8ab97ce6489b5d6" Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.731010 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi293e-account-delete-tb822"] Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.735692 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7xd65"] Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.740117 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7xd65"] Dec 05 11:34:43 crc kubenswrapper[4809]: I1205 11:34:43.755590 4809 scope.go:117] "RemoveContainer" containerID="1d84c3afeb7263e6079e6e47565f86fd4f1c3311f83150774d2707c65bc27b18" Dec 05 11:34:44 crc kubenswrapper[4809]: I1205 11:34:44.891898 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ae41780-4bf1-4a48-bf11-74dadb59e387" path="/var/lib/kubelet/pods/4ae41780-4bf1-4a48-bf11-74dadb59e387/volumes" Dec 05 11:34:44 crc kubenswrapper[4809]: I1205 11:34:44.893126 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" path="/var/lib/kubelet/pods/dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c/volumes" Dec 05 11:34:44 crc kubenswrapper[4809]: I1205 11:34:44.894266 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" path="/var/lib/kubelet/pods/e9c99d17-0a97-4f67-8295-68683c42a2c0/volumes" Dec 05 11:34:44 crc kubenswrapper[4809]: I1205 11:34:44.968963 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:44 crc kubenswrapper[4809]: I1205 11:34:44.969030 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:45 crc kubenswrapper[4809]: I1205 11:34:45.023221 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:48 crc kubenswrapper[4809]: I1205 11:34:48.872975 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:34:48 crc kubenswrapper[4809]: E1205 11:34:48.874001 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:34:55 crc kubenswrapper[4809]: I1205 11:34:55.045200 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vpwkx" Dec 05 11:34:55 crc kubenswrapper[4809]: I1205 11:34:55.151583 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vpwkx"] Dec 05 11:34:55 crc kubenswrapper[4809]: I1205 11:34:55.200369 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qspwd"] Dec 05 11:34:55 crc kubenswrapper[4809]: I1205 11:34:55.200811 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qspwd" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="registry-server" containerID="cri-o://5dcb8d720a19e31dd679c03e814aff742aae95e3b123a38e0cd881ceb3211818" gracePeriod=2 Dec 05 11:34:57 crc kubenswrapper[4809]: I1205 11:34:57.833605 4809 generic.go:334] "Generic (PLEG): container finished" podID="1299366a-d263-4417-aef3-eaa199c32e72" containerID="5dcb8d720a19e31dd679c03e814aff742aae95e3b123a38e0cd881ceb3211818" exitCode=0 Dec 05 11:34:57 crc kubenswrapper[4809]: I1205 11:34:57.833743 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qspwd" event={"ID":"1299366a-d263-4417-aef3-eaa199c32e72","Type":"ContainerDied","Data":"5dcb8d720a19e31dd679c03e814aff742aae95e3b123a38e0cd881ceb3211818"} Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.320465 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.479257 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-catalog-content\") pod \"1299366a-d263-4417-aef3-eaa199c32e72\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.479392 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-utilities\") pod \"1299366a-d263-4417-aef3-eaa199c32e72\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.479485 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wndbg\" (UniqueName: \"kubernetes.io/projected/1299366a-d263-4417-aef3-eaa199c32e72-kube-api-access-wndbg\") pod \"1299366a-d263-4417-aef3-eaa199c32e72\" (UID: \"1299366a-d263-4417-aef3-eaa199c32e72\") " Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.480111 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-utilities" (OuterVolumeSpecName: "utilities") pod "1299366a-d263-4417-aef3-eaa199c32e72" (UID: "1299366a-d263-4417-aef3-eaa199c32e72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.488950 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1299366a-d263-4417-aef3-eaa199c32e72-kube-api-access-wndbg" (OuterVolumeSpecName: "kube-api-access-wndbg") pod "1299366a-d263-4417-aef3-eaa199c32e72" (UID: "1299366a-d263-4417-aef3-eaa199c32e72"). InnerVolumeSpecName "kube-api-access-wndbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.496121 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1299366a-d263-4417-aef3-eaa199c32e72" (UID: "1299366a-d263-4417-aef3-eaa199c32e72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.581055 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.581084 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1299366a-d263-4417-aef3-eaa199c32e72-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.581094 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wndbg\" (UniqueName: \"kubernetes.io/projected/1299366a-d263-4417-aef3-eaa199c32e72-kube-api-access-wndbg\") on node \"crc\" DevicePath \"\"" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.844623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qspwd" event={"ID":"1299366a-d263-4417-aef3-eaa199c32e72","Type":"ContainerDied","Data":"1aeb51f4f79fade7cebb86c435a59de22221a814fbc2f431819f25e52079b36b"} Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.845618 4809 scope.go:117] "RemoveContainer" containerID="5dcb8d720a19e31dd679c03e814aff742aae95e3b123a38e0cd881ceb3211818" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.845884 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qspwd" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.885625 4809 scope.go:117] "RemoveContainer" containerID="7871fe02a77d8fa46af144fc6b2460c9c1123dea09d90a4d0addd66b1d9e3b8f" Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.894336 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qspwd"] Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.895867 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qspwd"] Dec 05 11:34:58 crc kubenswrapper[4809]: I1205 11:34:58.916491 4809 scope.go:117] "RemoveContainer" containerID="550c1b347633229b7d6ab14d55033e90af99ce4d5ce5e3e7aa046b8c18c31f8f" Dec 05 11:35:00 crc kubenswrapper[4809]: I1205 11:35:00.871865 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:35:00 crc kubenswrapper[4809]: E1205 11:35:00.872158 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:35:00 crc kubenswrapper[4809]: I1205 11:35:00.881800 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1299366a-d263-4417-aef3-eaa199c32e72" path="/var/lib/kubelet/pods/1299366a-d263-4417-aef3-eaa199c32e72/volumes" Dec 05 11:35:13 crc kubenswrapper[4809]: I1205 11:35:13.873187 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:35:13 crc kubenswrapper[4809]: E1205 11:35:13.874015 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.283091 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mvrck"] Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284073 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284107 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284129 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server-init" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284146 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server-init" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284174 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-updater" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284188 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-updater" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284208 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284220 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-server" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284241 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="cinder-scheduler" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284253 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="cinder-scheduler" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284270 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284283 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284301 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284313 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284333 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f68706-76e6-4ee3-b27d-bb092ab68a30" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284345 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f68706-76e6-4ee3-b27d-bb092ab68a30" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284365 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284378 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-server" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284400 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.284416 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.284444 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae41780-4bf1-4a48-bf11-74dadb59e387" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.289624 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae41780-4bf1-4a48-bf11-74dadb59e387" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.289701 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.289722 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.289782 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102c9fa9-707d-4918-9a5d-55d8d4c47c12" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.289800 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="102c9fa9-707d-4918-9a5d-55d8d4c47c12" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.289822 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="registry-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.289839 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="registry-server" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.289858 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="extract-utilities" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.289880 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="extract-utilities" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.289906 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="probe" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.289925 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="probe" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.289955 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-reaper" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.289974 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-reaper" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290004 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="extract-content" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290023 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="extract-content" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290045 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290062 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290091 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-expirer" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290108 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-expirer" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290139 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="extract-content" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290157 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="extract-content" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290179 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="rsync" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290195 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="rsync" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290213 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290230 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290266 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="swift-recon-cron" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290284 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="swift-recon-cron" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290312 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="registry-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290329 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="registry-server" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290367 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="extract-utilities" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290385 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="extract-utilities" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290412 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-updater" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290429 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-updater" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290455 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290473 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290493 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290509 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:35:16 crc kubenswrapper[4809]: E1205 11:35:16.290533 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.290549 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291031 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9c99d17-0a97-4f67-8295-68683c42a2c0" containerName="registry-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291075 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-expirer" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291102 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd461ce6-f0f2-4381-8cee-e7eeaf7ae41c" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291124 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovsdb-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291152 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="probe" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291172 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1299366a-d263-4417-aef3-eaa199c32e72" containerName="registry-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291202 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291224 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="11752f85-2262-477b-9920-bf5c13f45f20" containerName="ovs-vswitchd" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291251 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291279 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-reaper" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291304 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-updater" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291321 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f68706-76e6-4ee3-b27d-bb092ab68a30" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291344 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="102c9fa9-707d-4918-9a5d-55d8d4c47c12" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291374 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291398 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="rsync" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291426 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291454 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ae41780-4bf1-4a48-bf11-74dadb59e387" containerName="mariadb-account-delete" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291472 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d12a95a-49fa-4392-9a0d-9efd9aa28720" containerName="cinder-scheduler" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291496 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-updater" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291522 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-replicator" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291550 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="container-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291569 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291590 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="account-auditor" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291620 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="object-server" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.291685 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="58501ab1-18d7-45d9-98cb-a74e62db36fc" containerName="swift-recon-cron" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.294150 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.305736 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mvrck"] Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.308241 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-catalog-content\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.308360 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-utilities\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.308489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khcff\" (UniqueName: \"kubernetes.io/projected/e41572d0-24e1-4512-a6a0-664a5b3693d8-kube-api-access-khcff\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.409323 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-utilities\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.409828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khcff\" (UniqueName: \"kubernetes.io/projected/e41572d0-24e1-4512-a6a0-664a5b3693d8-kube-api-access-khcff\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.409984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-catalog-content\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.409907 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-utilities\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.410227 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-catalog-content\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.429990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khcff\" (UniqueName: \"kubernetes.io/projected/e41572d0-24e1-4512-a6a0-664a5b3693d8-kube-api-access-khcff\") pod \"certified-operators-mvrck\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:16 crc kubenswrapper[4809]: I1205 11:35:16.620200 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:17 crc kubenswrapper[4809]: I1205 11:35:17.115866 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mvrck"] Dec 05 11:35:18 crc kubenswrapper[4809]: I1205 11:35:18.046823 4809 generic.go:334] "Generic (PLEG): container finished" podID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerID="2a2518fb81140135ef3fbca50b64099e81bf8b20d005195234617e247cb81323" exitCode=0 Dec 05 11:35:18 crc kubenswrapper[4809]: I1205 11:35:18.047059 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrck" event={"ID":"e41572d0-24e1-4512-a6a0-664a5b3693d8","Type":"ContainerDied","Data":"2a2518fb81140135ef3fbca50b64099e81bf8b20d005195234617e247cb81323"} Dec 05 11:35:18 crc kubenswrapper[4809]: I1205 11:35:18.047313 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrck" event={"ID":"e41572d0-24e1-4512-a6a0-664a5b3693d8","Type":"ContainerStarted","Data":"6ab83ef7565b7efdba540f47e97aa53e9825461af1bdab8eccc156807dea51f9"} Dec 05 11:35:19 crc kubenswrapper[4809]: I1205 11:35:19.058710 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrck" event={"ID":"e41572d0-24e1-4512-a6a0-664a5b3693d8","Type":"ContainerStarted","Data":"3110a9a14bc5d54c4fc29e5638f8564c29bbbe6ecf728707516b523df6916490"} Dec 05 11:35:20 crc kubenswrapper[4809]: I1205 11:35:20.068474 4809 generic.go:334] "Generic (PLEG): container finished" podID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerID="3110a9a14bc5d54c4fc29e5638f8564c29bbbe6ecf728707516b523df6916490" exitCode=0 Dec 05 11:35:20 crc kubenswrapper[4809]: I1205 11:35:20.069305 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrck" event={"ID":"e41572d0-24e1-4512-a6a0-664a5b3693d8","Type":"ContainerDied","Data":"3110a9a14bc5d54c4fc29e5638f8564c29bbbe6ecf728707516b523df6916490"} Dec 05 11:35:21 crc kubenswrapper[4809]: I1205 11:35:21.080549 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrck" event={"ID":"e41572d0-24e1-4512-a6a0-664a5b3693d8","Type":"ContainerStarted","Data":"cf79eaf436cefadb5edfd85481f8ba13c23a75ad0d0b77207f6b0428c3b5a82d"} Dec 05 11:35:21 crc kubenswrapper[4809]: I1205 11:35:21.104045 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mvrck" podStartSLOduration=2.619816128 podStartE2EDuration="5.104018878s" podCreationTimestamp="2025-12-05 11:35:16 +0000 UTC" firstStartedPulling="2025-12-05 11:35:18.049150522 +0000 UTC m=+1613.440127090" lastFinishedPulling="2025-12-05 11:35:20.533353282 +0000 UTC m=+1615.924329840" observedRunningTime="2025-12-05 11:35:21.102948969 +0000 UTC m=+1616.493925557" watchObservedRunningTime="2025-12-05 11:35:21.104018878 +0000 UTC m=+1616.494995476" Dec 05 11:35:25 crc kubenswrapper[4809]: I1205 11:35:25.872009 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:35:25 crc kubenswrapper[4809]: E1205 11:35:25.872743 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:35:26 crc kubenswrapper[4809]: I1205 11:35:26.620413 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:26 crc kubenswrapper[4809]: I1205 11:35:26.620740 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:26 crc kubenswrapper[4809]: I1205 11:35:26.669398 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:27 crc kubenswrapper[4809]: I1205 11:35:27.212924 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:27 crc kubenswrapper[4809]: I1205 11:35:27.300889 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mvrck"] Dec 05 11:35:29 crc kubenswrapper[4809]: I1205 11:35:29.157107 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mvrck" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="registry-server" containerID="cri-o://cf79eaf436cefadb5edfd85481f8ba13c23a75ad0d0b77207f6b0428c3b5a82d" gracePeriod=2 Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.170997 4809 generic.go:334] "Generic (PLEG): container finished" podID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerID="cf79eaf436cefadb5edfd85481f8ba13c23a75ad0d0b77207f6b0428c3b5a82d" exitCode=0 Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.171082 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrck" event={"ID":"e41572d0-24e1-4512-a6a0-664a5b3693d8","Type":"ContainerDied","Data":"cf79eaf436cefadb5edfd85481f8ba13c23a75ad0d0b77207f6b0428c3b5a82d"} Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.487618 4809 scope.go:117] "RemoveContainer" containerID="ab993a42330cf1af383e0abbeeb94faa4345af10b7cf650185f7e021bb2a7d93" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.535248 4809 scope.go:117] "RemoveContainer" containerID="f395b9608a6f94d2a59890234bff67c7844f4946ecc0490217665ccdc20f313b" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.577271 4809 scope.go:117] "RemoveContainer" containerID="b5512e72b384254d31928502d46c73cced07ecb562b80bd83398a34c2cdbb8ef" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.622852 4809 scope.go:117] "RemoveContainer" containerID="4692bcfe7142bc2738fd2b0178e41f14055eeab3c3ae7f26d1ad3477b9ee970c" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.643801 4809 scope.go:117] "RemoveContainer" containerID="4501f7a7e855fa971938992a20e039bcf63b75bfe786a79b79d5ebde13b721ba" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.682013 4809 scope.go:117] "RemoveContainer" containerID="ca17fb3b9f7c4840c3062e70874e1e17ae9c14450b3a2c90ff5e4142c3b55b40" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.718226 4809 scope.go:117] "RemoveContainer" containerID="7387e0246accba106019ec8b535f729502b108f72a360c3a7ec7e0a7c8203dbd" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.743300 4809 scope.go:117] "RemoveContainer" containerID="bebae10222675f498e6edbdca8558f09457b6b61eba7576c606fad86d41f948e" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.775798 4809 scope.go:117] "RemoveContainer" containerID="e1ef014dab365de170de8bcc18bc67b0c2a056e8f731e71b7a9f0c26d47e7122" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.795620 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.951622 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khcff\" (UniqueName: \"kubernetes.io/projected/e41572d0-24e1-4512-a6a0-664a5b3693d8-kube-api-access-khcff\") pod \"e41572d0-24e1-4512-a6a0-664a5b3693d8\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.951917 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-utilities\") pod \"e41572d0-24e1-4512-a6a0-664a5b3693d8\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.952026 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-catalog-content\") pod \"e41572d0-24e1-4512-a6a0-664a5b3693d8\" (UID: \"e41572d0-24e1-4512-a6a0-664a5b3693d8\") " Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.952818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-utilities" (OuterVolumeSpecName: "utilities") pod "e41572d0-24e1-4512-a6a0-664a5b3693d8" (UID: "e41572d0-24e1-4512-a6a0-664a5b3693d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:35:30 crc kubenswrapper[4809]: I1205 11:35:30.957902 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e41572d0-24e1-4512-a6a0-664a5b3693d8-kube-api-access-khcff" (OuterVolumeSpecName: "kube-api-access-khcff") pod "e41572d0-24e1-4512-a6a0-664a5b3693d8" (UID: "e41572d0-24e1-4512-a6a0-664a5b3693d8"). InnerVolumeSpecName "kube-api-access-khcff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.008230 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e41572d0-24e1-4512-a6a0-664a5b3693d8" (UID: "e41572d0-24e1-4512-a6a0-664a5b3693d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.053892 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khcff\" (UniqueName: \"kubernetes.io/projected/e41572d0-24e1-4512-a6a0-664a5b3693d8-kube-api-access-khcff\") on node \"crc\" DevicePath \"\"" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.053918 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.053928 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e41572d0-24e1-4512-a6a0-664a5b3693d8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.183661 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mvrck" event={"ID":"e41572d0-24e1-4512-a6a0-664a5b3693d8","Type":"ContainerDied","Data":"6ab83ef7565b7efdba540f47e97aa53e9825461af1bdab8eccc156807dea51f9"} Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.183729 4809 scope.go:117] "RemoveContainer" containerID="cf79eaf436cefadb5edfd85481f8ba13c23a75ad0d0b77207f6b0428c3b5a82d" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.183776 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mvrck" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.203114 4809 scope.go:117] "RemoveContainer" containerID="3110a9a14bc5d54c4fc29e5638f8564c29bbbe6ecf728707516b523df6916490" Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.226992 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mvrck"] Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.232999 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mvrck"] Dec 05 11:35:31 crc kubenswrapper[4809]: I1205 11:35:31.240621 4809 scope.go:117] "RemoveContainer" containerID="2a2518fb81140135ef3fbca50b64099e81bf8b20d005195234617e247cb81323" Dec 05 11:35:32 crc kubenswrapper[4809]: I1205 11:35:32.883872 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" path="/var/lib/kubelet/pods/e41572d0-24e1-4512-a6a0-664a5b3693d8/volumes" Dec 05 11:35:36 crc kubenswrapper[4809]: I1205 11:35:36.873004 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:35:36 crc kubenswrapper[4809]: E1205 11:35:36.873925 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:35:50 crc kubenswrapper[4809]: I1205 11:35:50.873313 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:35:50 crc kubenswrapper[4809]: E1205 11:35:50.874279 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:36:03 crc kubenswrapper[4809]: I1205 11:36:03.872820 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:36:03 crc kubenswrapper[4809]: E1205 11:36:03.874146 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:36:15 crc kubenswrapper[4809]: I1205 11:36:15.873789 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:36:15 crc kubenswrapper[4809]: E1205 11:36:15.874789 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:36:28 crc kubenswrapper[4809]: I1205 11:36:28.873436 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:36:28 crc kubenswrapper[4809]: E1205 11:36:28.874598 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.047471 4809 scope.go:117] "RemoveContainer" containerID="e4918c7f89901cb6bc80954b1fdf3a854bdcb6bfe697fc6fc6ad89234ed88f5d" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.084021 4809 scope.go:117] "RemoveContainer" containerID="a81cb4726f7d91229efe101694be71b159de9cb7acbee23af76ed162d4c42eab" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.123208 4809 scope.go:117] "RemoveContainer" containerID="e66ce562a684ce2f7c9a5765f371c11805fe1afdfbe8c023db95e39baf7fd55d" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.177757 4809 scope.go:117] "RemoveContainer" containerID="63cb23ada3d99ae3c4ef0234a641edf1c1ba2827cee6b00495704fb452a0053e" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.212812 4809 scope.go:117] "RemoveContainer" containerID="aad7adfa2eb8eca2b71c201df7cb8a7e5b0a169a22eecd2c354e23760da7454b" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.243950 4809 scope.go:117] "RemoveContainer" containerID="62e9456b1210de31be32cf45c13c879bc91ee026ca8abe1d1406a82b2c2bc9fb" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.271243 4809 scope.go:117] "RemoveContainer" containerID="1ad92427b11dd57c6028550f7cc630d7917be3d0d72e445aa75f2d9dd8a6ac7b" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.296972 4809 scope.go:117] "RemoveContainer" containerID="e0b33876d9e64c7fd99f042a76d398b7db13a4e6ce06d131130aa7bed7b10815" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.317778 4809 scope.go:117] "RemoveContainer" containerID="f615d4be35da2629c2d18907ba7e6166ab3ef4c48f963d7e1f9a7707bd8f6399" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.336973 4809 scope.go:117] "RemoveContainer" containerID="b2f329660098f5d914595479a66fbd5e0b73b1633354e05daff2ff0d359489c9" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.354864 4809 scope.go:117] "RemoveContainer" containerID="f74b9b84031262b69b17ddf856c579346742b85b32806bca97cf70ab2d4209da" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.378944 4809 scope.go:117] "RemoveContainer" containerID="32e6b77cca37dcdbc635d76201fe078ccada3173763f886d6471946dd5311728" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.411613 4809 scope.go:117] "RemoveContainer" containerID="f6b8afa57cce5400518588e578205bb1cb1cdecff67726b8da835d625cf70a77" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.432987 4809 scope.go:117] "RemoveContainer" containerID="4124bbc4606e9f85acf01d756fb541e366fb21a27e987393909ff3ed0e222a54" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.464753 4809 scope.go:117] "RemoveContainer" containerID="d0a4c11aa9fc03b76e0161912fbe85f1f20894ffb26f11d0e3172a02fc7f3658" Dec 05 11:36:31 crc kubenswrapper[4809]: I1205 11:36:31.490502 4809 scope.go:117] "RemoveContainer" containerID="0be7f476d8685657cf0bb6c394823faf079fd3c3df299ee0d38d4ac3075cf7ae" Dec 05 11:36:42 crc kubenswrapper[4809]: I1205 11:36:42.873140 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:36:42 crc kubenswrapper[4809]: E1205 11:36:42.874258 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:36:54 crc kubenswrapper[4809]: I1205 11:36:54.875788 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:36:54 crc kubenswrapper[4809]: E1205 11:36:54.876514 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:37:08 crc kubenswrapper[4809]: I1205 11:37:08.873046 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:37:08 crc kubenswrapper[4809]: E1205 11:37:08.873936 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:37:22 crc kubenswrapper[4809]: I1205 11:37:22.872102 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:37:22 crc kubenswrapper[4809]: E1205 11:37:22.873128 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.703429 4809 scope.go:117] "RemoveContainer" containerID="2c109d5ee7420de221d8af2681db967c2c239aa7aa67e96c11e16f774976f28d" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.737770 4809 scope.go:117] "RemoveContainer" containerID="624371613c304b126b4f3ff33d0fd6cb21766812caf78100b3abf43bd31dedf7" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.830377 4809 scope.go:117] "RemoveContainer" containerID="0c7d459938d65d4994155346009222c7ccaf5872fe4d76d2c6e4cb10a6a0e64d" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.863580 4809 scope.go:117] "RemoveContainer" containerID="fd7d9a735a223752ce2da66b2a6e5377eebafb13bb894d79878a71ea6f44a4a7" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.915502 4809 scope.go:117] "RemoveContainer" containerID="fb2ba8a351288b4e9e91e7356e1bca3ff3d3d9546db734be03367086d01d5f39" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.937845 4809 scope.go:117] "RemoveContainer" containerID="258c92cf388669a38f82f3d4a6a139c7c0d61b4d6ea7930faec41f22a47d2809" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.956080 4809 scope.go:117] "RemoveContainer" containerID="4770b253a022abba8402b06255199087a163db712ada152cea565a2565eb0ffc" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.977029 4809 scope.go:117] "RemoveContainer" containerID="9a0cf769fac3a0d40431db27c3e061f9a49d341f34de1e9d46c4e22c316f1114" Dec 05 11:37:31 crc kubenswrapper[4809]: I1205 11:37:31.998540 4809 scope.go:117] "RemoveContainer" containerID="eddaf5222111460d42f2358e99f093a8dd969f342da92ec3db6de0ad73712f04" Dec 05 11:37:32 crc kubenswrapper[4809]: I1205 11:37:32.022622 4809 scope.go:117] "RemoveContainer" containerID="a3de22e1ff3f7898f9fee8f16d5df757205e16ec84e88652001b2d777b01c877" Dec 05 11:37:32 crc kubenswrapper[4809]: I1205 11:37:32.037830 4809 scope.go:117] "RemoveContainer" containerID="068867c80f55e49b4b5a5f4b24b296d240d64a45c30b6fee625ac91d520dc067" Dec 05 11:37:32 crc kubenswrapper[4809]: I1205 11:37:32.065843 4809 scope.go:117] "RemoveContainer" containerID="6dc5d0d87115d603eb8dd2e73cec49003fe6ab7534b7c9f6a2b38814ea7d4fcb" Dec 05 11:37:32 crc kubenswrapper[4809]: I1205 11:37:32.087570 4809 scope.go:117] "RemoveContainer" containerID="b355c3d21b69548d3439df18096cd1fdb75313f1444a93680f88ddda6c381b12" Dec 05 11:37:34 crc kubenswrapper[4809]: I1205 11:37:34.880280 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:37:34 crc kubenswrapper[4809]: E1205 11:37:34.881091 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:37:49 crc kubenswrapper[4809]: I1205 11:37:49.872212 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:37:49 crc kubenswrapper[4809]: E1205 11:37:49.872792 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:38:00 crc kubenswrapper[4809]: I1205 11:38:00.872900 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:38:00 crc kubenswrapper[4809]: E1205 11:38:00.873824 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:38:12 crc kubenswrapper[4809]: I1205 11:38:12.872512 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:38:12 crc kubenswrapper[4809]: E1205 11:38:12.873242 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:38:24 crc kubenswrapper[4809]: I1205 11:38:24.880958 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:38:24 crc kubenswrapper[4809]: E1205 11:38:24.882048 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:38:32 crc kubenswrapper[4809]: I1205 11:38:32.247365 4809 scope.go:117] "RemoveContainer" containerID="9af9fd7f8d169447279e6fa340bbf6d3969e9c9fff1059f5d537b02471d23b2d" Dec 05 11:38:32 crc kubenswrapper[4809]: I1205 11:38:32.295898 4809 scope.go:117] "RemoveContainer" containerID="227b5a95db28ac44992a513775b179933f3871062f8c30164d38b84a81d68994" Dec 05 11:38:32 crc kubenswrapper[4809]: I1205 11:38:32.334975 4809 scope.go:117] "RemoveContainer" containerID="57697b760919bef59126e13516e74ae1bb694c94addef30e58466b3dc060be8a" Dec 05 11:38:32 crc kubenswrapper[4809]: I1205 11:38:32.356412 4809 scope.go:117] "RemoveContainer" containerID="d66ed99c8c65211097d15a89bde5605afdc64db5d12e59bb72da9739aeaa7f01" Dec 05 11:38:32 crc kubenswrapper[4809]: I1205 11:38:32.399794 4809 scope.go:117] "RemoveContainer" containerID="894c6277b8486291c59a23e96d6969fc19f092834afc5ce802d6b8dc9d69bcc6" Dec 05 11:38:39 crc kubenswrapper[4809]: I1205 11:38:39.872442 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:38:39 crc kubenswrapper[4809]: E1205 11:38:39.873331 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:38:52 crc kubenswrapper[4809]: I1205 11:38:52.872580 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:38:52 crc kubenswrapper[4809]: E1205 11:38:52.873905 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:39:07 crc kubenswrapper[4809]: I1205 11:39:07.871993 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:39:07 crc kubenswrapper[4809]: E1205 11:39:07.875460 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:39:18 crc kubenswrapper[4809]: I1205 11:39:18.872726 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:39:19 crc kubenswrapper[4809]: I1205 11:39:19.429317 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"82c6870bff0754b116a9b679069b0f3e0e769fd9c5f2b1f5d84bf4edf3d8fab8"} Dec 05 11:39:32 crc kubenswrapper[4809]: I1205 11:39:32.496444 4809 scope.go:117] "RemoveContainer" containerID="fc8f6fbd0d6af548b3ab28ac3220e10e5ec059642c2119b7cb235f5f82975b77" Dec 05 11:39:32 crc kubenswrapper[4809]: I1205 11:39:32.526949 4809 scope.go:117] "RemoveContainer" containerID="99c7a952c43dee727c2d091290d87f6671731e8e5972b24248e47525e3da6a3a" Dec 05 11:39:32 crc kubenswrapper[4809]: I1205 11:39:32.564664 4809 scope.go:117] "RemoveContainer" containerID="1fa8ff72d7fcc2c42aaea2405f43852074e7d8c0caf9f1b4cb78b7401add51b2" Dec 05 11:39:32 crc kubenswrapper[4809]: I1205 11:39:32.587186 4809 scope.go:117] "RemoveContainer" containerID="cee2aebc7c4c395069fa4d92acc3a5a78a920e62eb51e07edd4caaa0df65aaca" Dec 05 11:40:32 crc kubenswrapper[4809]: I1205 11:40:32.742189 4809 scope.go:117] "RemoveContainer" containerID="7db9bc6598923a51269cb0d779ffcff5d06cf7c742e40da04fae4c8d368b202b" Dec 05 11:40:32 crc kubenswrapper[4809]: I1205 11:40:32.769242 4809 scope.go:117] "RemoveContainer" containerID="9fcbf16b8475f5eccd9c0a78f2fdd8efeeeff59735de6888340ba66f89160e1f" Dec 05 11:40:32 crc kubenswrapper[4809]: I1205 11:40:32.801339 4809 scope.go:117] "RemoveContainer" containerID="03aa6fe669afc3afb549e6a085150e79d198307149b8f8ad7ff8518cad774a58" Dec 05 11:41:44 crc kubenswrapper[4809]: I1205 11:41:44.046300 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:41:44 crc kubenswrapper[4809]: I1205 11:41:44.046817 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:42:14 crc kubenswrapper[4809]: I1205 11:42:14.046326 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:42:14 crc kubenswrapper[4809]: I1205 11:42:14.046836 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.047293 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.047877 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.047922 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.048656 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"82c6870bff0754b116a9b679069b0f3e0e769fd9c5f2b1f5d84bf4edf3d8fab8"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.048775 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://82c6870bff0754b116a9b679069b0f3e0e769fd9c5f2b1f5d84bf4edf3d8fab8" gracePeriod=600 Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.282318 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="82c6870bff0754b116a9b679069b0f3e0e769fd9c5f2b1f5d84bf4edf3d8fab8" exitCode=0 Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.282368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"82c6870bff0754b116a9b679069b0f3e0e769fd9c5f2b1f5d84bf4edf3d8fab8"} Dec 05 11:42:44 crc kubenswrapper[4809]: I1205 11:42:44.282422 4809 scope.go:117] "RemoveContainer" containerID="4e92a9ffa44f25286150c2544bf56e2e094d131b320e6b975b0f5f206388a0a9" Dec 05 11:42:45 crc kubenswrapper[4809]: I1205 11:42:45.295357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2"} Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.650166 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vhkpz"] Dec 05 11:44:35 crc kubenswrapper[4809]: E1205 11:44:35.651147 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="registry-server" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.651166 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="registry-server" Dec 05 11:44:35 crc kubenswrapper[4809]: E1205 11:44:35.651203 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="extract-utilities" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.651212 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="extract-utilities" Dec 05 11:44:35 crc kubenswrapper[4809]: E1205 11:44:35.651228 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="extract-content" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.651236 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="extract-content" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.651414 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e41572d0-24e1-4512-a6a0-664a5b3693d8" containerName="registry-server" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.652702 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.664259 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhkpz"] Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.750683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx4hm\" (UniqueName: \"kubernetes.io/projected/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-kube-api-access-xx4hm\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.750753 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-catalog-content\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.750851 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-utilities\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.852654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-utilities\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.852731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx4hm\" (UniqueName: \"kubernetes.io/projected/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-kube-api-access-xx4hm\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.852752 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-catalog-content\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.853209 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-utilities\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.853215 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-catalog-content\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.873816 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx4hm\" (UniqueName: \"kubernetes.io/projected/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-kube-api-access-xx4hm\") pod \"redhat-marketplace-vhkpz\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:35 crc kubenswrapper[4809]: I1205 11:44:35.972424 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:36 crc kubenswrapper[4809]: I1205 11:44:36.397825 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhkpz"] Dec 05 11:44:36 crc kubenswrapper[4809]: W1205 11:44:36.405337 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3351e492_e6b2_41e2_8c3f_c3cd5ef63c16.slice/crio-13ed59890cdef76b2746f009a86aa522b0ae0f7178800969d343046ba5d20059 WatchSource:0}: Error finding container 13ed59890cdef76b2746f009a86aa522b0ae0f7178800969d343046ba5d20059: Status 404 returned error can't find the container with id 13ed59890cdef76b2746f009a86aa522b0ae0f7178800969d343046ba5d20059 Dec 05 11:44:37 crc kubenswrapper[4809]: I1205 11:44:37.229835 4809 generic.go:334] "Generic (PLEG): container finished" podID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerID="b02cb1ffe852a94b8c4a5e712cd67b0630169837dec6a000663488084050e9de" exitCode=0 Dec 05 11:44:37 crc kubenswrapper[4809]: I1205 11:44:37.229898 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhkpz" event={"ID":"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16","Type":"ContainerDied","Data":"b02cb1ffe852a94b8c4a5e712cd67b0630169837dec6a000663488084050e9de"} Dec 05 11:44:37 crc kubenswrapper[4809]: I1205 11:44:37.230197 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhkpz" event={"ID":"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16","Type":"ContainerStarted","Data":"13ed59890cdef76b2746f009a86aa522b0ae0f7178800969d343046ba5d20059"} Dec 05 11:44:37 crc kubenswrapper[4809]: I1205 11:44:37.233435 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.240045 4809 generic.go:334] "Generic (PLEG): container finished" podID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerID="3bc30f5fe24a54523a635b9b9739b75348aded890e3600213587d5bb0c7e1a01" exitCode=0 Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.240097 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhkpz" event={"ID":"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16","Type":"ContainerDied","Data":"3bc30f5fe24a54523a635b9b9739b75348aded890e3600213587d5bb0c7e1a01"} Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.429161 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qq4gv"] Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.430877 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.447371 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qq4gv"] Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.602318 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-catalog-content\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.602439 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6b9b\" (UniqueName: \"kubernetes.io/projected/83f831aa-b8d0-41ba-9b67-1c1353c738ea-kube-api-access-v6b9b\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.602536 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-utilities\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.703728 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-utilities\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.703805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-catalog-content\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.703843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6b9b\" (UniqueName: \"kubernetes.io/projected/83f831aa-b8d0-41ba-9b67-1c1353c738ea-kube-api-access-v6b9b\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.704487 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-utilities\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.704974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-catalog-content\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.724573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6b9b\" (UniqueName: \"kubernetes.io/projected/83f831aa-b8d0-41ba-9b67-1c1353c738ea-kube-api-access-v6b9b\") pod \"redhat-operators-qq4gv\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:38 crc kubenswrapper[4809]: I1205 11:44:38.787584 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:39 crc kubenswrapper[4809]: I1205 11:44:39.238220 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qq4gv"] Dec 05 11:44:39 crc kubenswrapper[4809]: I1205 11:44:39.256638 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhkpz" event={"ID":"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16","Type":"ContainerStarted","Data":"2118fb889a9db262059576e02e033f68e9679c0acc445ad391e235ef4ee8b4db"} Dec 05 11:44:39 crc kubenswrapper[4809]: I1205 11:44:39.283442 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vhkpz" podStartSLOduration=2.86069403 podStartE2EDuration="4.283419332s" podCreationTimestamp="2025-12-05 11:44:35 +0000 UTC" firstStartedPulling="2025-12-05 11:44:37.232942876 +0000 UTC m=+2172.623919464" lastFinishedPulling="2025-12-05 11:44:38.655668198 +0000 UTC m=+2174.046644766" observedRunningTime="2025-12-05 11:44:39.278666273 +0000 UTC m=+2174.669642831" watchObservedRunningTime="2025-12-05 11:44:39.283419332 +0000 UTC m=+2174.674395890" Dec 05 11:44:40 crc kubenswrapper[4809]: I1205 11:44:40.263183 4809 generic.go:334] "Generic (PLEG): container finished" podID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerID="bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f" exitCode=0 Dec 05 11:44:40 crc kubenswrapper[4809]: I1205 11:44:40.263252 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qq4gv" event={"ID":"83f831aa-b8d0-41ba-9b67-1c1353c738ea","Type":"ContainerDied","Data":"bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f"} Dec 05 11:44:40 crc kubenswrapper[4809]: I1205 11:44:40.263537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qq4gv" event={"ID":"83f831aa-b8d0-41ba-9b67-1c1353c738ea","Type":"ContainerStarted","Data":"603e968a6c654e99f49bcf3bf8b40ad1f6a3c6f152f03ae6e36aca303c99ae71"} Dec 05 11:44:41 crc kubenswrapper[4809]: I1205 11:44:41.273925 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qq4gv" event={"ID":"83f831aa-b8d0-41ba-9b67-1c1353c738ea","Type":"ContainerStarted","Data":"1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5"} Dec 05 11:44:42 crc kubenswrapper[4809]: I1205 11:44:42.310423 4809 generic.go:334] "Generic (PLEG): container finished" podID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerID="1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5" exitCode=0 Dec 05 11:44:42 crc kubenswrapper[4809]: I1205 11:44:42.310490 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qq4gv" event={"ID":"83f831aa-b8d0-41ba-9b67-1c1353c738ea","Type":"ContainerDied","Data":"1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5"} Dec 05 11:44:43 crc kubenswrapper[4809]: I1205 11:44:43.319995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qq4gv" event={"ID":"83f831aa-b8d0-41ba-9b67-1c1353c738ea","Type":"ContainerStarted","Data":"66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051"} Dec 05 11:44:43 crc kubenswrapper[4809]: I1205 11:44:43.343734 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qq4gv" podStartSLOduration=2.900773092 podStartE2EDuration="5.343716151s" podCreationTimestamp="2025-12-05 11:44:38 +0000 UTC" firstStartedPulling="2025-12-05 11:44:40.26459095 +0000 UTC m=+2175.655567508" lastFinishedPulling="2025-12-05 11:44:42.707534009 +0000 UTC m=+2178.098510567" observedRunningTime="2025-12-05 11:44:43.340618277 +0000 UTC m=+2178.731594845" watchObservedRunningTime="2025-12-05 11:44:43.343716151 +0000 UTC m=+2178.734692709" Dec 05 11:44:44 crc kubenswrapper[4809]: I1205 11:44:44.046348 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:44:44 crc kubenswrapper[4809]: I1205 11:44:44.046401 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:44:45 crc kubenswrapper[4809]: I1205 11:44:45.973336 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:45 crc kubenswrapper[4809]: I1205 11:44:45.973705 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:46 crc kubenswrapper[4809]: I1205 11:44:46.018350 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:46 crc kubenswrapper[4809]: I1205 11:44:46.426232 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:47 crc kubenswrapper[4809]: I1205 11:44:47.234283 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhkpz"] Dec 05 11:44:48 crc kubenswrapper[4809]: I1205 11:44:48.373666 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vhkpz" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="registry-server" containerID="cri-o://2118fb889a9db262059576e02e033f68e9679c0acc445ad391e235ef4ee8b4db" gracePeriod=2 Dec 05 11:44:48 crc kubenswrapper[4809]: I1205 11:44:48.788394 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:48 crc kubenswrapper[4809]: I1205 11:44:48.788803 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:48 crc kubenswrapper[4809]: I1205 11:44:48.840077 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:49 crc kubenswrapper[4809]: I1205 11:44:49.386712 4809 generic.go:334] "Generic (PLEG): container finished" podID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerID="2118fb889a9db262059576e02e033f68e9679c0acc445ad391e235ef4ee8b4db" exitCode=0 Dec 05 11:44:49 crc kubenswrapper[4809]: I1205 11:44:49.386828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhkpz" event={"ID":"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16","Type":"ContainerDied","Data":"2118fb889a9db262059576e02e033f68e9679c0acc445ad391e235ef4ee8b4db"} Dec 05 11:44:49 crc kubenswrapper[4809]: I1205 11:44:49.468467 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:49 crc kubenswrapper[4809]: I1205 11:44:49.962550 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.073344 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-catalog-content\") pod \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.073444 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-utilities\") pod \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.073581 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx4hm\" (UniqueName: \"kubernetes.io/projected/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-kube-api-access-xx4hm\") pod \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\" (UID: \"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16\") " Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.075023 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-utilities" (OuterVolumeSpecName: "utilities") pod "3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" (UID: "3351e492-e6b2-41e2-8c3f-c3cd5ef63c16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.075528 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.082859 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-kube-api-access-xx4hm" (OuterVolumeSpecName: "kube-api-access-xx4hm") pod "3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" (UID: "3351e492-e6b2-41e2-8c3f-c3cd5ef63c16"). InnerVolumeSpecName "kube-api-access-xx4hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.098983 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" (UID: "3351e492-e6b2-41e2-8c3f-c3cd5ef63c16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.177424 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx4hm\" (UniqueName: \"kubernetes.io/projected/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-kube-api-access-xx4hm\") on node \"crc\" DevicePath \"\"" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.177471 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.401962 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhkpz" event={"ID":"3351e492-e6b2-41e2-8c3f-c3cd5ef63c16","Type":"ContainerDied","Data":"13ed59890cdef76b2746f009a86aa522b0ae0f7178800969d343046ba5d20059"} Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.402015 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhkpz" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.402049 4809 scope.go:117] "RemoveContainer" containerID="2118fb889a9db262059576e02e033f68e9679c0acc445ad391e235ef4ee8b4db" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.443582 4809 scope.go:117] "RemoveContainer" containerID="3bc30f5fe24a54523a635b9b9739b75348aded890e3600213587d5bb0c7e1a01" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.456225 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhkpz"] Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.463887 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhkpz"] Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.483807 4809 scope.go:117] "RemoveContainer" containerID="b02cb1ffe852a94b8c4a5e712cd67b0630169837dec6a000663488084050e9de" Dec 05 11:44:50 crc kubenswrapper[4809]: I1205 11:44:50.887947 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" path="/var/lib/kubelet/pods/3351e492-e6b2-41e2-8c3f-c3cd5ef63c16/volumes" Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.230907 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qq4gv"] Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.413875 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qq4gv" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="registry-server" containerID="cri-o://66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051" gracePeriod=2 Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.895181 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.913929 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-utilities\") pod \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.914236 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6b9b\" (UniqueName: \"kubernetes.io/projected/83f831aa-b8d0-41ba-9b67-1c1353c738ea-kube-api-access-v6b9b\") pod \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.914269 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-catalog-content\") pod \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\" (UID: \"83f831aa-b8d0-41ba-9b67-1c1353c738ea\") " Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.915065 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-utilities" (OuterVolumeSpecName: "utilities") pod "83f831aa-b8d0-41ba-9b67-1c1353c738ea" (UID: "83f831aa-b8d0-41ba-9b67-1c1353c738ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:44:51 crc kubenswrapper[4809]: I1205 11:44:51.920087 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f831aa-b8d0-41ba-9b67-1c1353c738ea-kube-api-access-v6b9b" (OuterVolumeSpecName: "kube-api-access-v6b9b") pod "83f831aa-b8d0-41ba-9b67-1c1353c738ea" (UID: "83f831aa-b8d0-41ba-9b67-1c1353c738ea"). InnerVolumeSpecName "kube-api-access-v6b9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.015447 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6b9b\" (UniqueName: \"kubernetes.io/projected/83f831aa-b8d0-41ba-9b67-1c1353c738ea-kube-api-access-v6b9b\") on node \"crc\" DevicePath \"\"" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.015482 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.423004 4809 generic.go:334] "Generic (PLEG): container finished" podID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerID="66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051" exitCode=0 Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.423078 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qq4gv" event={"ID":"83f831aa-b8d0-41ba-9b67-1c1353c738ea","Type":"ContainerDied","Data":"66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051"} Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.423266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qq4gv" event={"ID":"83f831aa-b8d0-41ba-9b67-1c1353c738ea","Type":"ContainerDied","Data":"603e968a6c654e99f49bcf3bf8b40ad1f6a3c6f152f03ae6e36aca303c99ae71"} Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.423104 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qq4gv" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.423285 4809 scope.go:117] "RemoveContainer" containerID="66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.443795 4809 scope.go:117] "RemoveContainer" containerID="1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.467025 4809 scope.go:117] "RemoveContainer" containerID="bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.486521 4809 scope.go:117] "RemoveContainer" containerID="66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051" Dec 05 11:44:52 crc kubenswrapper[4809]: E1205 11:44:52.487118 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051\": container with ID starting with 66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051 not found: ID does not exist" containerID="66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.487162 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051"} err="failed to get container status \"66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051\": rpc error: code = NotFound desc = could not find container \"66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051\": container with ID starting with 66669c782cecd24c01315a4e26375f54b333812277a4931d01b517a29df3b051 not found: ID does not exist" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.487191 4809 scope.go:117] "RemoveContainer" containerID="1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5" Dec 05 11:44:52 crc kubenswrapper[4809]: E1205 11:44:52.487722 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5\": container with ID starting with 1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5 not found: ID does not exist" containerID="1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.487771 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5"} err="failed to get container status \"1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5\": rpc error: code = NotFound desc = could not find container \"1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5\": container with ID starting with 1e658a266eefca5ba4b5c38f320a56fc942bcb6770400789f045fec7bad3a7d5 not found: ID does not exist" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.487802 4809 scope.go:117] "RemoveContainer" containerID="bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f" Dec 05 11:44:52 crc kubenswrapper[4809]: E1205 11:44:52.488228 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f\": container with ID starting with bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f not found: ID does not exist" containerID="bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.488252 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f"} err="failed to get container status \"bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f\": rpc error: code = NotFound desc = could not find container \"bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f\": container with ID starting with bb5b62fbab17efc12548b50799cfe47001f462936ae076ef5b40a6737e4cfb4f not found: ID does not exist" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.736297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83f831aa-b8d0-41ba-9b67-1c1353c738ea" (UID: "83f831aa-b8d0-41ba-9b67-1c1353c738ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:44:52 crc kubenswrapper[4809]: I1205 11:44:52.826501 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f831aa-b8d0-41ba-9b67-1c1353c738ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:44:53 crc kubenswrapper[4809]: I1205 11:44:53.047075 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qq4gv"] Dec 05 11:44:53 crc kubenswrapper[4809]: I1205 11:44:53.058155 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qq4gv"] Dec 05 11:44:54 crc kubenswrapper[4809]: I1205 11:44:54.886626 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" path="/var/lib/kubelet/pods/83f831aa-b8d0-41ba-9b67-1c1353c738ea/volumes" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.148899 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm"] Dec 05 11:45:00 crc kubenswrapper[4809]: E1205 11:45:00.150951 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="registry-server" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.151066 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="registry-server" Dec 05 11:45:00 crc kubenswrapper[4809]: E1205 11:45:00.151176 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="registry-server" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.151259 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="registry-server" Dec 05 11:45:00 crc kubenswrapper[4809]: E1205 11:45:00.151346 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="extract-utilities" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.151431 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="extract-utilities" Dec 05 11:45:00 crc kubenswrapper[4809]: E1205 11:45:00.151524 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="extract-content" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.151602 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="extract-content" Dec 05 11:45:00 crc kubenswrapper[4809]: E1205 11:45:00.151726 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="extract-content" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.151812 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="extract-content" Dec 05 11:45:00 crc kubenswrapper[4809]: E1205 11:45:00.151894 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="extract-utilities" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.151974 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="extract-utilities" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.152231 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f831aa-b8d0-41ba-9b67-1c1353c738ea" containerName="registry-server" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.152318 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3351e492-e6b2-41e2-8c3f-c3cd5ef63c16" containerName="registry-server" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.153020 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.155590 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.155743 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.168041 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm"] Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.343294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8r8w\" (UniqueName: \"kubernetes.io/projected/7b9cf7e7-71b9-47df-ac05-09bc6406360f-kube-api-access-c8r8w\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.343385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b9cf7e7-71b9-47df-ac05-09bc6406360f-secret-volume\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.343482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b9cf7e7-71b9-47df-ac05-09bc6406360f-config-volume\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.444472 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8r8w\" (UniqueName: \"kubernetes.io/projected/7b9cf7e7-71b9-47df-ac05-09bc6406360f-kube-api-access-c8r8w\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.444574 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b9cf7e7-71b9-47df-ac05-09bc6406360f-secret-volume\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.444697 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b9cf7e7-71b9-47df-ac05-09bc6406360f-config-volume\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.446042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b9cf7e7-71b9-47df-ac05-09bc6406360f-config-volume\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.453408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b9cf7e7-71b9-47df-ac05-09bc6406360f-secret-volume\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.463080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8r8w\" (UniqueName: \"kubernetes.io/projected/7b9cf7e7-71b9-47df-ac05-09bc6406360f-kube-api-access-c8r8w\") pod \"collect-profiles-29415585-j55qm\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.474018 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:00 crc kubenswrapper[4809]: I1205 11:45:00.934241 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm"] Dec 05 11:45:01 crc kubenswrapper[4809]: I1205 11:45:01.920474 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" event={"ID":"7b9cf7e7-71b9-47df-ac05-09bc6406360f","Type":"ContainerStarted","Data":"3e724a30a7424decd8547720e5c046c2320cde2c7effbbf23b2fbd4e596f058f"} Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.815845 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4b4b6"] Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.818056 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.833344 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4b4b6"] Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.930040 4809 generic.go:334] "Generic (PLEG): container finished" podID="7b9cf7e7-71b9-47df-ac05-09bc6406360f" containerID="7afd2813ee8499e8cf1972b63b4cbbe6b72f0d8268faf1322758dff9e62250f2" exitCode=0 Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.930101 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" event={"ID":"7b9cf7e7-71b9-47df-ac05-09bc6406360f","Type":"ContainerDied","Data":"7afd2813ee8499e8cf1972b63b4cbbe6b72f0d8268faf1322758dff9e62250f2"} Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.982676 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-utilities\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.982808 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-catalog-content\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:02 crc kubenswrapper[4809]: I1205 11:45:02.982985 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8qkg\" (UniqueName: \"kubernetes.io/projected/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-kube-api-access-g8qkg\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.084934 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-utilities\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.085008 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-catalog-content\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.085109 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8qkg\" (UniqueName: \"kubernetes.io/projected/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-kube-api-access-g8qkg\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.085489 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-utilities\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.087598 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-catalog-content\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.112392 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8qkg\" (UniqueName: \"kubernetes.io/projected/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-kube-api-access-g8qkg\") pod \"community-operators-4b4b6\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.145908 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.622775 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4b4b6"] Dec 05 11:45:03 crc kubenswrapper[4809]: W1205 11:45:03.625277 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode66a0bdd_cf78_4f26_a481_16dbe33e6f2d.slice/crio-efacf46627755186d6d9e70031c9c8fcaab0da608cf9c073764e56d5016f8dc0 WatchSource:0}: Error finding container efacf46627755186d6d9e70031c9c8fcaab0da608cf9c073764e56d5016f8dc0: Status 404 returned error can't find the container with id efacf46627755186d6d9e70031c9c8fcaab0da608cf9c073764e56d5016f8dc0 Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.945060 4809 generic.go:334] "Generic (PLEG): container finished" podID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerID="4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e" exitCode=0 Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.945168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4b4b6" event={"ID":"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d","Type":"ContainerDied","Data":"4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e"} Dec 05 11:45:03 crc kubenswrapper[4809]: I1205 11:45:03.945212 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4b4b6" event={"ID":"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d","Type":"ContainerStarted","Data":"efacf46627755186d6d9e70031c9c8fcaab0da608cf9c073764e56d5016f8dc0"} Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.304092 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.503794 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b9cf7e7-71b9-47df-ac05-09bc6406360f-config-volume\") pod \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.503840 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8r8w\" (UniqueName: \"kubernetes.io/projected/7b9cf7e7-71b9-47df-ac05-09bc6406360f-kube-api-access-c8r8w\") pod \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.503931 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b9cf7e7-71b9-47df-ac05-09bc6406360f-secret-volume\") pod \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\" (UID: \"7b9cf7e7-71b9-47df-ac05-09bc6406360f\") " Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.505583 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b9cf7e7-71b9-47df-ac05-09bc6406360f-config-volume" (OuterVolumeSpecName: "config-volume") pod "7b9cf7e7-71b9-47df-ac05-09bc6406360f" (UID: "7b9cf7e7-71b9-47df-ac05-09bc6406360f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.511605 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b9cf7e7-71b9-47df-ac05-09bc6406360f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7b9cf7e7-71b9-47df-ac05-09bc6406360f" (UID: "7b9cf7e7-71b9-47df-ac05-09bc6406360f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.512844 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b9cf7e7-71b9-47df-ac05-09bc6406360f-kube-api-access-c8r8w" (OuterVolumeSpecName: "kube-api-access-c8r8w") pod "7b9cf7e7-71b9-47df-ac05-09bc6406360f" (UID: "7b9cf7e7-71b9-47df-ac05-09bc6406360f"). InnerVolumeSpecName "kube-api-access-c8r8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.605465 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b9cf7e7-71b9-47df-ac05-09bc6406360f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.605823 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b9cf7e7-71b9-47df-ac05-09bc6406360f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.605836 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8r8w\" (UniqueName: \"kubernetes.io/projected/7b9cf7e7-71b9-47df-ac05-09bc6406360f-kube-api-access-c8r8w\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.954494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" event={"ID":"7b9cf7e7-71b9-47df-ac05-09bc6406360f","Type":"ContainerDied","Data":"3e724a30a7424decd8547720e5c046c2320cde2c7effbbf23b2fbd4e596f058f"} Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.954577 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm" Dec 05 11:45:04 crc kubenswrapper[4809]: I1205 11:45:04.954598 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e724a30a7424decd8547720e5c046c2320cde2c7effbbf23b2fbd4e596f058f" Dec 05 11:45:05 crc kubenswrapper[4809]: I1205 11:45:05.377908 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8"] Dec 05 11:45:05 crc kubenswrapper[4809]: I1205 11:45:05.383245 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-ccjd8"] Dec 05 11:45:05 crc kubenswrapper[4809]: I1205 11:45:05.966297 4809 generic.go:334] "Generic (PLEG): container finished" podID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerID="b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d" exitCode=0 Dec 05 11:45:05 crc kubenswrapper[4809]: I1205 11:45:05.966368 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4b4b6" event={"ID":"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d","Type":"ContainerDied","Data":"b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d"} Dec 05 11:45:06 crc kubenswrapper[4809]: I1205 11:45:06.880904 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e9748a-1d74-4b39-9bf0-5c94dda32218" path="/var/lib/kubelet/pods/e7e9748a-1d74-4b39-9bf0-5c94dda32218/volumes" Dec 05 11:45:06 crc kubenswrapper[4809]: I1205 11:45:06.976829 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4b4b6" event={"ID":"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d","Type":"ContainerStarted","Data":"271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8"} Dec 05 11:45:13 crc kubenswrapper[4809]: I1205 11:45:13.146860 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:13 crc kubenswrapper[4809]: I1205 11:45:13.147593 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:13 crc kubenswrapper[4809]: I1205 11:45:13.210351 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:13 crc kubenswrapper[4809]: I1205 11:45:13.227892 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4b4b6" podStartSLOduration=8.769221375 podStartE2EDuration="11.227872755s" podCreationTimestamp="2025-12-05 11:45:02 +0000 UTC" firstStartedPulling="2025-12-05 11:45:03.946783805 +0000 UTC m=+2199.337760363" lastFinishedPulling="2025-12-05 11:45:06.405435185 +0000 UTC m=+2201.796411743" observedRunningTime="2025-12-05 11:45:07.014049223 +0000 UTC m=+2202.405025781" watchObservedRunningTime="2025-12-05 11:45:13.227872755 +0000 UTC m=+2208.618849313" Dec 05 11:45:14 crc kubenswrapper[4809]: I1205 11:45:14.046518 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:45:14 crc kubenswrapper[4809]: I1205 11:45:14.047022 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:45:14 crc kubenswrapper[4809]: I1205 11:45:14.119071 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:14 crc kubenswrapper[4809]: I1205 11:45:14.188529 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4b4b6"] Dec 05 11:45:16 crc kubenswrapper[4809]: I1205 11:45:16.064345 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4b4b6" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="registry-server" containerID="cri-o://271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8" gracePeriod=2 Dec 05 11:45:16 crc kubenswrapper[4809]: I1205 11:45:16.993988 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.073294 4809 generic.go:334] "Generic (PLEG): container finished" podID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerID="271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8" exitCode=0 Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.073343 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4b4b6" event={"ID":"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d","Type":"ContainerDied","Data":"271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8"} Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.073369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4b4b6" event={"ID":"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d","Type":"ContainerDied","Data":"efacf46627755186d6d9e70031c9c8fcaab0da608cf9c073764e56d5016f8dc0"} Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.073392 4809 scope.go:117] "RemoveContainer" containerID="271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.073460 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4b4b6" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.091086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-catalog-content\") pod \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.091142 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8qkg\" (UniqueName: \"kubernetes.io/projected/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-kube-api-access-g8qkg\") pod \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.091174 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-utilities\") pod \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\" (UID: \"e66a0bdd-cf78-4f26-a481-16dbe33e6f2d\") " Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.092425 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-utilities" (OuterVolumeSpecName: "utilities") pod "e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" (UID: "e66a0bdd-cf78-4f26-a481-16dbe33e6f2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.097974 4809 scope.go:117] "RemoveContainer" containerID="b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.099909 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-kube-api-access-g8qkg" (OuterVolumeSpecName: "kube-api-access-g8qkg") pod "e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" (UID: "e66a0bdd-cf78-4f26-a481-16dbe33e6f2d"). InnerVolumeSpecName "kube-api-access-g8qkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.132705 4809 scope.go:117] "RemoveContainer" containerID="4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.153839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" (UID: "e66a0bdd-cf78-4f26-a481-16dbe33e6f2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.155331 4809 scope.go:117] "RemoveContainer" containerID="271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8" Dec 05 11:45:17 crc kubenswrapper[4809]: E1205 11:45:17.155946 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8\": container with ID starting with 271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8 not found: ID does not exist" containerID="271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.155998 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8"} err="failed to get container status \"271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8\": rpc error: code = NotFound desc = could not find container \"271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8\": container with ID starting with 271540afebf35fbc15f8be0f24a396ab89daa5821e28645df0403a372c3539a8 not found: ID does not exist" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.156030 4809 scope.go:117] "RemoveContainer" containerID="b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d" Dec 05 11:45:17 crc kubenswrapper[4809]: E1205 11:45:17.156426 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d\": container with ID starting with b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d not found: ID does not exist" containerID="b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.156475 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d"} err="failed to get container status \"b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d\": rpc error: code = NotFound desc = could not find container \"b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d\": container with ID starting with b6f5a87866064e4b07f395d672e13630a9f613e867e495ac652cd0fda7db7a1d not found: ID does not exist" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.156494 4809 scope.go:117] "RemoveContainer" containerID="4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e" Dec 05 11:45:17 crc kubenswrapper[4809]: E1205 11:45:17.157053 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e\": container with ID starting with 4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e not found: ID does not exist" containerID="4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.157097 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e"} err="failed to get container status \"4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e\": rpc error: code = NotFound desc = could not find container \"4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e\": container with ID starting with 4b299edfd08fad1e8001e5ddda21de5424244387d27f704be42c726a90cf603e not found: ID does not exist" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.192786 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.192823 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8qkg\" (UniqueName: \"kubernetes.io/projected/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-kube-api-access-g8qkg\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.192836 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.410246 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4b4b6"] Dec 05 11:45:17 crc kubenswrapper[4809]: I1205 11:45:17.418509 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4b4b6"] Dec 05 11:45:18 crc kubenswrapper[4809]: I1205 11:45:18.886677 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" path="/var/lib/kubelet/pods/e66a0bdd-cf78-4f26-a481-16dbe33e6f2d/volumes" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.744571 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v72dk"] Dec 05 11:45:22 crc kubenswrapper[4809]: E1205 11:45:22.746919 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9cf7e7-71b9-47df-ac05-09bc6406360f" containerName="collect-profiles" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.746950 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9cf7e7-71b9-47df-ac05-09bc6406360f" containerName="collect-profiles" Dec 05 11:45:22 crc kubenswrapper[4809]: E1205 11:45:22.746973 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="extract-content" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.746985 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="extract-content" Dec 05 11:45:22 crc kubenswrapper[4809]: E1205 11:45:22.747028 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="registry-server" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.747040 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="registry-server" Dec 05 11:45:22 crc kubenswrapper[4809]: E1205 11:45:22.747059 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="extract-utilities" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.747071 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="extract-utilities" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.747349 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b9cf7e7-71b9-47df-ac05-09bc6406360f" containerName="collect-profiles" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.747396 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e66a0bdd-cf78-4f26-a481-16dbe33e6f2d" containerName="registry-server" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.749342 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.763042 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v72dk"] Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.880243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-catalog-content\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.880317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-utilities\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.880734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw4hs\" (UniqueName: \"kubernetes.io/projected/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-kube-api-access-tw4hs\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.982579 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-catalog-content\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.982695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-utilities\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.982871 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw4hs\" (UniqueName: \"kubernetes.io/projected/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-kube-api-access-tw4hs\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.983480 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-catalog-content\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:22 crc kubenswrapper[4809]: I1205 11:45:22.983961 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-utilities\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:23 crc kubenswrapper[4809]: I1205 11:45:23.007533 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw4hs\" (UniqueName: \"kubernetes.io/projected/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-kube-api-access-tw4hs\") pod \"certified-operators-v72dk\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:23 crc kubenswrapper[4809]: I1205 11:45:23.083887 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:23 crc kubenswrapper[4809]: I1205 11:45:23.416229 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v72dk"] Dec 05 11:45:24 crc kubenswrapper[4809]: I1205 11:45:24.145806 4809 generic.go:334] "Generic (PLEG): container finished" podID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerID="5887cdc8c823e0ec67d95e7219d1991cbe5eed5264e83768eb12d17402e71dc2" exitCode=0 Dec 05 11:45:24 crc kubenswrapper[4809]: I1205 11:45:24.146097 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v72dk" event={"ID":"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd","Type":"ContainerDied","Data":"5887cdc8c823e0ec67d95e7219d1991cbe5eed5264e83768eb12d17402e71dc2"} Dec 05 11:45:24 crc kubenswrapper[4809]: I1205 11:45:24.146121 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v72dk" event={"ID":"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd","Type":"ContainerStarted","Data":"9a463a36c44c551a6ab5760759d6fcbff70b87d530a7c47a7917db6f6fe1007f"} Dec 05 11:45:25 crc kubenswrapper[4809]: I1205 11:45:25.155437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v72dk" event={"ID":"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd","Type":"ContainerStarted","Data":"b70d0ec29314b2876c0fe74b5b6e662faa63815b3a8271d5a08ce7ab83090f27"} Dec 05 11:45:26 crc kubenswrapper[4809]: I1205 11:45:26.170291 4809 generic.go:334] "Generic (PLEG): container finished" podID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerID="b70d0ec29314b2876c0fe74b5b6e662faa63815b3a8271d5a08ce7ab83090f27" exitCode=0 Dec 05 11:45:26 crc kubenswrapper[4809]: I1205 11:45:26.170545 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v72dk" event={"ID":"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd","Type":"ContainerDied","Data":"b70d0ec29314b2876c0fe74b5b6e662faa63815b3a8271d5a08ce7ab83090f27"} Dec 05 11:45:27 crc kubenswrapper[4809]: I1205 11:45:27.181461 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v72dk" event={"ID":"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd","Type":"ContainerStarted","Data":"c1a41637afcb14eff26e9d6e19eab0db8852e0a888c4438936c2791edd6da8f9"} Dec 05 11:45:27 crc kubenswrapper[4809]: I1205 11:45:27.205581 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v72dk" podStartSLOduration=2.811868344 podStartE2EDuration="5.205558238s" podCreationTimestamp="2025-12-05 11:45:22 +0000 UTC" firstStartedPulling="2025-12-05 11:45:24.147280545 +0000 UTC m=+2219.538257113" lastFinishedPulling="2025-12-05 11:45:26.540970439 +0000 UTC m=+2221.931947007" observedRunningTime="2025-12-05 11:45:27.203308937 +0000 UTC m=+2222.594285505" watchObservedRunningTime="2025-12-05 11:45:27.205558238 +0000 UTC m=+2222.596534806" Dec 05 11:45:32 crc kubenswrapper[4809]: I1205 11:45:32.951226 4809 scope.go:117] "RemoveContainer" containerID="20e225726cb8690fef73f681e053f73bf9cdd1d6c0813a3e790395c81fac2e47" Dec 05 11:45:33 crc kubenswrapper[4809]: I1205 11:45:33.084520 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:33 crc kubenswrapper[4809]: I1205 11:45:33.084579 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:33 crc kubenswrapper[4809]: I1205 11:45:33.161208 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:33 crc kubenswrapper[4809]: I1205 11:45:33.293787 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:37 crc kubenswrapper[4809]: I1205 11:45:37.734596 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v72dk"] Dec 05 11:45:37 crc kubenswrapper[4809]: I1205 11:45:37.735700 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v72dk" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="registry-server" containerID="cri-o://c1a41637afcb14eff26e9d6e19eab0db8852e0a888c4438936c2791edd6da8f9" gracePeriod=2 Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.288333 4809 generic.go:334] "Generic (PLEG): container finished" podID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerID="c1a41637afcb14eff26e9d6e19eab0db8852e0a888c4438936c2791edd6da8f9" exitCode=0 Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.288416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v72dk" event={"ID":"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd","Type":"ContainerDied","Data":"c1a41637afcb14eff26e9d6e19eab0db8852e0a888c4438936c2791edd6da8f9"} Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.361745 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.467353 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-utilities\") pod \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.467441 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw4hs\" (UniqueName: \"kubernetes.io/projected/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-kube-api-access-tw4hs\") pod \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.467531 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-catalog-content\") pod \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\" (UID: \"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd\") " Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.469186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-utilities" (OuterVolumeSpecName: "utilities") pod "1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" (UID: "1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.475826 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-kube-api-access-tw4hs" (OuterVolumeSpecName: "kube-api-access-tw4hs") pod "1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" (UID: "1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd"). InnerVolumeSpecName "kube-api-access-tw4hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.532856 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" (UID: "1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.570162 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw4hs\" (UniqueName: \"kubernetes.io/projected/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-kube-api-access-tw4hs\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.570200 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:39 crc kubenswrapper[4809]: I1205 11:45:39.570210 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.303121 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v72dk" event={"ID":"1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd","Type":"ContainerDied","Data":"9a463a36c44c551a6ab5760759d6fcbff70b87d530a7c47a7917db6f6fe1007f"} Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.303221 4809 scope.go:117] "RemoveContainer" containerID="c1a41637afcb14eff26e9d6e19eab0db8852e0a888c4438936c2791edd6da8f9" Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.303251 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v72dk" Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.345482 4809 scope.go:117] "RemoveContainer" containerID="b70d0ec29314b2876c0fe74b5b6e662faa63815b3a8271d5a08ce7ab83090f27" Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.367190 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v72dk"] Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.373923 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v72dk"] Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.387991 4809 scope.go:117] "RemoveContainer" containerID="5887cdc8c823e0ec67d95e7219d1991cbe5eed5264e83768eb12d17402e71dc2" Dec 05 11:45:40 crc kubenswrapper[4809]: I1205 11:45:40.893937 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" path="/var/lib/kubelet/pods/1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd/volumes" Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.047017 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.047365 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.047415 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.048064 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.048147 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" gracePeriod=600 Dec 05 11:45:44 crc kubenswrapper[4809]: E1205 11:45:44.166513 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.348245 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" exitCode=0 Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.348291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2"} Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.348330 4809 scope.go:117] "RemoveContainer" containerID="82c6870bff0754b116a9b679069b0f3e0e769fd9c5f2b1f5d84bf4edf3d8fab8" Dec 05 11:45:44 crc kubenswrapper[4809]: I1205 11:45:44.349094 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:45:44 crc kubenswrapper[4809]: E1205 11:45:44.349451 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:45:55 crc kubenswrapper[4809]: I1205 11:45:55.872885 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:45:55 crc kubenswrapper[4809]: E1205 11:45:55.874281 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:46:07 crc kubenswrapper[4809]: I1205 11:46:07.873053 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:46:07 crc kubenswrapper[4809]: E1205 11:46:07.874334 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:46:19 crc kubenswrapper[4809]: I1205 11:46:19.872000 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:46:19 crc kubenswrapper[4809]: E1205 11:46:19.872872 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:46:31 crc kubenswrapper[4809]: I1205 11:46:31.872971 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:46:31 crc kubenswrapper[4809]: E1205 11:46:31.874059 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:46:46 crc kubenswrapper[4809]: I1205 11:46:46.872412 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:46:46 crc kubenswrapper[4809]: E1205 11:46:46.874561 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:47:01 crc kubenswrapper[4809]: I1205 11:47:01.872266 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:47:01 crc kubenswrapper[4809]: E1205 11:47:01.873137 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:47:15 crc kubenswrapper[4809]: I1205 11:47:15.872564 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:47:15 crc kubenswrapper[4809]: E1205 11:47:15.873819 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:47:26 crc kubenswrapper[4809]: I1205 11:47:26.872851 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:47:26 crc kubenswrapper[4809]: E1205 11:47:26.873798 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:47:39 crc kubenswrapper[4809]: I1205 11:47:39.872865 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:47:39 crc kubenswrapper[4809]: E1205 11:47:39.874025 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:47:52 crc kubenswrapper[4809]: I1205 11:47:52.873174 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:47:52 crc kubenswrapper[4809]: E1205 11:47:52.874766 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:48:03 crc kubenswrapper[4809]: I1205 11:48:03.872060 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:48:03 crc kubenswrapper[4809]: E1205 11:48:03.872948 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:48:15 crc kubenswrapper[4809]: I1205 11:48:15.872609 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:48:15 crc kubenswrapper[4809]: E1205 11:48:15.873830 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:48:26 crc kubenswrapper[4809]: I1205 11:48:26.871938 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:48:26 crc kubenswrapper[4809]: E1205 11:48:26.872673 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:48:41 crc kubenswrapper[4809]: I1205 11:48:41.872979 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:48:41 crc kubenswrapper[4809]: E1205 11:48:41.874166 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:48:52 crc kubenswrapper[4809]: I1205 11:48:52.872995 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:48:52 crc kubenswrapper[4809]: E1205 11:48:52.875993 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:49:05 crc kubenswrapper[4809]: I1205 11:49:05.872397 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:49:05 crc kubenswrapper[4809]: E1205 11:49:05.873038 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:49:19 crc kubenswrapper[4809]: I1205 11:49:19.872212 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:49:19 crc kubenswrapper[4809]: E1205 11:49:19.873106 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:49:30 crc kubenswrapper[4809]: I1205 11:49:30.871889 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:49:30 crc kubenswrapper[4809]: E1205 11:49:30.872620 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:49:43 crc kubenswrapper[4809]: I1205 11:49:43.872131 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:49:43 crc kubenswrapper[4809]: E1205 11:49:43.872857 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:49:54 crc kubenswrapper[4809]: I1205 11:49:54.877528 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:49:54 crc kubenswrapper[4809]: E1205 11:49:54.878298 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:50:08 crc kubenswrapper[4809]: I1205 11:50:08.872777 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:50:08 crc kubenswrapper[4809]: E1205 11:50:08.873590 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:50:19 crc kubenswrapper[4809]: I1205 11:50:19.873219 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:50:19 crc kubenswrapper[4809]: E1205 11:50:19.874288 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:50:30 crc kubenswrapper[4809]: I1205 11:50:30.872971 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:50:30 crc kubenswrapper[4809]: E1205 11:50:30.873964 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:50:41 crc kubenswrapper[4809]: I1205 11:50:41.872800 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:50:41 crc kubenswrapper[4809]: E1205 11:50:41.874085 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:50:55 crc kubenswrapper[4809]: I1205 11:50:55.872284 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:50:56 crc kubenswrapper[4809]: I1205 11:50:56.368067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"6da7f8756648d6aa0d5dcf32827c2417d82a2103f968e117591720c658b834b2"} Dec 05 11:53:14 crc kubenswrapper[4809]: I1205 11:53:14.047049 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:53:14 crc kubenswrapper[4809]: I1205 11:53:14.047854 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:53:44 crc kubenswrapper[4809]: I1205 11:53:44.047742 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:53:44 crc kubenswrapper[4809]: I1205 11:53:44.048503 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:54:14 crc kubenswrapper[4809]: I1205 11:54:14.046866 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:54:14 crc kubenswrapper[4809]: I1205 11:54:14.047311 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:54:14 crc kubenswrapper[4809]: I1205 11:54:14.047355 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:54:14 crc kubenswrapper[4809]: I1205 11:54:14.047956 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6da7f8756648d6aa0d5dcf32827c2417d82a2103f968e117591720c658b834b2"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:54:14 crc kubenswrapper[4809]: I1205 11:54:14.048007 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://6da7f8756648d6aa0d5dcf32827c2417d82a2103f968e117591720c658b834b2" gracePeriod=600 Dec 05 11:54:15 crc kubenswrapper[4809]: I1205 11:54:15.072265 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="6da7f8756648d6aa0d5dcf32827c2417d82a2103f968e117591720c658b834b2" exitCode=0 Dec 05 11:54:15 crc kubenswrapper[4809]: I1205 11:54:15.072346 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"6da7f8756648d6aa0d5dcf32827c2417d82a2103f968e117591720c658b834b2"} Dec 05 11:54:15 crc kubenswrapper[4809]: I1205 11:54:15.072852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da"} Dec 05 11:54:15 crc kubenswrapper[4809]: I1205 11:54:15.072880 4809 scope.go:117] "RemoveContainer" containerID="dfe77627074116e4e77425aca76d9d5760e8cb6430d2570bc8af93534b9874c2" Dec 05 11:55:08 crc kubenswrapper[4809]: I1205 11:55:08.944301 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nqtlx"] Dec 05 11:55:08 crc kubenswrapper[4809]: E1205 11:55:08.945064 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="registry-server" Dec 05 11:55:08 crc kubenswrapper[4809]: I1205 11:55:08.945076 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="registry-server" Dec 05 11:55:08 crc kubenswrapper[4809]: E1205 11:55:08.945099 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="extract-content" Dec 05 11:55:08 crc kubenswrapper[4809]: I1205 11:55:08.945105 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="extract-content" Dec 05 11:55:08 crc kubenswrapper[4809]: E1205 11:55:08.945125 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="extract-utilities" Dec 05 11:55:08 crc kubenswrapper[4809]: I1205 11:55:08.945133 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="extract-utilities" Dec 05 11:55:08 crc kubenswrapper[4809]: I1205 11:55:08.945281 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e44a6bf-d8d0-4316-9fa0-601c8deb5dfd" containerName="registry-server" Dec 05 11:55:08 crc kubenswrapper[4809]: I1205 11:55:08.948174 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:08 crc kubenswrapper[4809]: I1205 11:55:08.954259 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqtlx"] Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.001174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhvcs\" (UniqueName: \"kubernetes.io/projected/e38052d2-7a63-47d5-ba61-005a8bdc32d1-kube-api-access-hhvcs\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.001266 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-utilities\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.001297 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-catalog-content\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.102983 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhvcs\" (UniqueName: \"kubernetes.io/projected/e38052d2-7a63-47d5-ba61-005a8bdc32d1-kube-api-access-hhvcs\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.103319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-utilities\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.103445 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-catalog-content\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.103981 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-utilities\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.104011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-catalog-content\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.121697 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhvcs\" (UniqueName: \"kubernetes.io/projected/e38052d2-7a63-47d5-ba61-005a8bdc32d1-kube-api-access-hhvcs\") pod \"redhat-marketplace-nqtlx\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.270168 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:09 crc kubenswrapper[4809]: I1205 11:55:09.713667 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqtlx"] Dec 05 11:55:10 crc kubenswrapper[4809]: I1205 11:55:10.503842 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqtlx" event={"ID":"e38052d2-7a63-47d5-ba61-005a8bdc32d1","Type":"ContainerStarted","Data":"55290b80acf1d581c361e6bb735402e10435ce702b8cd6f8370d1a27855c2bd3"} Dec 05 11:55:11 crc kubenswrapper[4809]: I1205 11:55:11.511938 4809 generic.go:334] "Generic (PLEG): container finished" podID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerID="7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502" exitCode=0 Dec 05 11:55:11 crc kubenswrapper[4809]: I1205 11:55:11.512002 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqtlx" event={"ID":"e38052d2-7a63-47d5-ba61-005a8bdc32d1","Type":"ContainerDied","Data":"7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502"} Dec 05 11:55:11 crc kubenswrapper[4809]: I1205 11:55:11.513723 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:55:14 crc kubenswrapper[4809]: I1205 11:55:14.535912 4809 generic.go:334] "Generic (PLEG): container finished" podID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerID="62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869" exitCode=0 Dec 05 11:55:14 crc kubenswrapper[4809]: I1205 11:55:14.535990 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqtlx" event={"ID":"e38052d2-7a63-47d5-ba61-005a8bdc32d1","Type":"ContainerDied","Data":"62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869"} Dec 05 11:55:15 crc kubenswrapper[4809]: I1205 11:55:15.555233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqtlx" event={"ID":"e38052d2-7a63-47d5-ba61-005a8bdc32d1","Type":"ContainerStarted","Data":"a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab"} Dec 05 11:55:15 crc kubenswrapper[4809]: I1205 11:55:15.579177 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nqtlx" podStartSLOduration=4.127205524 podStartE2EDuration="7.57915246s" podCreationTimestamp="2025-12-05 11:55:08 +0000 UTC" firstStartedPulling="2025-12-05 11:55:11.513381623 +0000 UTC m=+2806.904358181" lastFinishedPulling="2025-12-05 11:55:14.965328559 +0000 UTC m=+2810.356305117" observedRunningTime="2025-12-05 11:55:15.572023369 +0000 UTC m=+2810.962999947" watchObservedRunningTime="2025-12-05 11:55:15.57915246 +0000 UTC m=+2810.970129038" Dec 05 11:55:19 crc kubenswrapper[4809]: I1205 11:55:19.270868 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:19 crc kubenswrapper[4809]: I1205 11:55:19.271622 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:19 crc kubenswrapper[4809]: I1205 11:55:19.315056 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:29 crc kubenswrapper[4809]: I1205 11:55:29.316407 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:29 crc kubenswrapper[4809]: I1205 11:55:29.361379 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqtlx"] Dec 05 11:55:29 crc kubenswrapper[4809]: I1205 11:55:29.666649 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nqtlx" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="registry-server" containerID="cri-o://a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab" gracePeriod=2 Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.150868 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.213113 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhvcs\" (UniqueName: \"kubernetes.io/projected/e38052d2-7a63-47d5-ba61-005a8bdc32d1-kube-api-access-hhvcs\") pod \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.213200 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-utilities\") pod \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.213360 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-catalog-content\") pod \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\" (UID: \"e38052d2-7a63-47d5-ba61-005a8bdc32d1\") " Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.214297 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-utilities" (OuterVolumeSpecName: "utilities") pod "e38052d2-7a63-47d5-ba61-005a8bdc32d1" (UID: "e38052d2-7a63-47d5-ba61-005a8bdc32d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.221826 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e38052d2-7a63-47d5-ba61-005a8bdc32d1-kube-api-access-hhvcs" (OuterVolumeSpecName: "kube-api-access-hhvcs") pod "e38052d2-7a63-47d5-ba61-005a8bdc32d1" (UID: "e38052d2-7a63-47d5-ba61-005a8bdc32d1"). InnerVolumeSpecName "kube-api-access-hhvcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.251534 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e38052d2-7a63-47d5-ba61-005a8bdc32d1" (UID: "e38052d2-7a63-47d5-ba61-005a8bdc32d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.315707 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.315780 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhvcs\" (UniqueName: \"kubernetes.io/projected/e38052d2-7a63-47d5-ba61-005a8bdc32d1-kube-api-access-hhvcs\") on node \"crc\" DevicePath \"\"" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.315794 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e38052d2-7a63-47d5-ba61-005a8bdc32d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.682102 4809 generic.go:334] "Generic (PLEG): container finished" podID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerID="a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab" exitCode=0 Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.682194 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqtlx" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.682209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqtlx" event={"ID":"e38052d2-7a63-47d5-ba61-005a8bdc32d1","Type":"ContainerDied","Data":"a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab"} Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.682748 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqtlx" event={"ID":"e38052d2-7a63-47d5-ba61-005a8bdc32d1","Type":"ContainerDied","Data":"55290b80acf1d581c361e6bb735402e10435ce702b8cd6f8370d1a27855c2bd3"} Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.682773 4809 scope.go:117] "RemoveContainer" containerID="a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.713037 4809 scope.go:117] "RemoveContainer" containerID="62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.721415 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqtlx"] Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.728504 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqtlx"] Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.754958 4809 scope.go:117] "RemoveContainer" containerID="7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.773036 4809 scope.go:117] "RemoveContainer" containerID="a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab" Dec 05 11:55:30 crc kubenswrapper[4809]: E1205 11:55:30.773832 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab\": container with ID starting with a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab not found: ID does not exist" containerID="a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.773936 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab"} err="failed to get container status \"a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab\": rpc error: code = NotFound desc = could not find container \"a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab\": container with ID starting with a4efbd0caa5c775ef9de6c34f1ede2fde717ab8b785127fa078477a43c1a71ab not found: ID does not exist" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.774028 4809 scope.go:117] "RemoveContainer" containerID="62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869" Dec 05 11:55:30 crc kubenswrapper[4809]: E1205 11:55:30.776675 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869\": container with ID starting with 62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869 not found: ID does not exist" containerID="62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.776762 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869"} err="failed to get container status \"62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869\": rpc error: code = NotFound desc = could not find container \"62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869\": container with ID starting with 62bb43a88ebf47b8a82acd85112f37684165046e511661dc5bf120c07789b869 not found: ID does not exist" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.776819 4809 scope.go:117] "RemoveContainer" containerID="7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502" Dec 05 11:55:30 crc kubenswrapper[4809]: E1205 11:55:30.777405 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502\": container with ID starting with 7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502 not found: ID does not exist" containerID="7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.777441 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502"} err="failed to get container status \"7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502\": rpc error: code = NotFound desc = could not find container \"7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502\": container with ID starting with 7215b2a41da820363ef4092aa463a272b75cf6a3dc3863845ec4bbccf3479502 not found: ID does not exist" Dec 05 11:55:30 crc kubenswrapper[4809]: I1205 11:55:30.881611 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" path="/var/lib/kubelet/pods/e38052d2-7a63-47d5-ba61-005a8bdc32d1/volumes" Dec 05 11:55:32 crc kubenswrapper[4809]: I1205 11:55:32.968057 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cks67"] Dec 05 11:55:32 crc kubenswrapper[4809]: E1205 11:55:32.968864 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="extract-content" Dec 05 11:55:32 crc kubenswrapper[4809]: I1205 11:55:32.968887 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="extract-content" Dec 05 11:55:32 crc kubenswrapper[4809]: E1205 11:55:32.968930 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="extract-utilities" Dec 05 11:55:32 crc kubenswrapper[4809]: I1205 11:55:32.968942 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="extract-utilities" Dec 05 11:55:32 crc kubenswrapper[4809]: E1205 11:55:32.968969 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="registry-server" Dec 05 11:55:32 crc kubenswrapper[4809]: I1205 11:55:32.968979 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="registry-server" Dec 05 11:55:32 crc kubenswrapper[4809]: I1205 11:55:32.969202 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e38052d2-7a63-47d5-ba61-005a8bdc32d1" containerName="registry-server" Dec 05 11:55:32 crc kubenswrapper[4809]: I1205 11:55:32.970721 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:32 crc kubenswrapper[4809]: I1205 11:55:32.981422 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cks67"] Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.057672 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gktrk\" (UniqueName: \"kubernetes.io/projected/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-kube-api-access-gktrk\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.057748 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-utilities\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.057802 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-catalog-content\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.159027 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gktrk\" (UniqueName: \"kubernetes.io/projected/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-kube-api-access-gktrk\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.159089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-utilities\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.159135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-catalog-content\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.159737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-catalog-content\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.159877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-utilities\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.180398 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gktrk\" (UniqueName: \"kubernetes.io/projected/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-kube-api-access-gktrk\") pod \"community-operators-cks67\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.299872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:33 crc kubenswrapper[4809]: I1205 11:55:33.839731 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cks67"] Dec 05 11:55:34 crc kubenswrapper[4809]: I1205 11:55:34.710524 4809 generic.go:334] "Generic (PLEG): container finished" podID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerID="737a7057f31c7eaec3ba1416f5ecf68ca0e721f79855b854efd8e6ab71d25d9e" exitCode=0 Dec 05 11:55:34 crc kubenswrapper[4809]: I1205 11:55:34.710599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cks67" event={"ID":"b2a801ec-eed1-4b07-bdb2-1654c69ffb98","Type":"ContainerDied","Data":"737a7057f31c7eaec3ba1416f5ecf68ca0e721f79855b854efd8e6ab71d25d9e"} Dec 05 11:55:34 crc kubenswrapper[4809]: I1205 11:55:34.711775 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cks67" event={"ID":"b2a801ec-eed1-4b07-bdb2-1654c69ffb98","Type":"ContainerStarted","Data":"ba6f30eba66bf39d324086a5b40ec4384ae499dda9fed9321f1e5ffc934abbf1"} Dec 05 11:55:36 crc kubenswrapper[4809]: I1205 11:55:36.740186 4809 generic.go:334] "Generic (PLEG): container finished" podID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerID="fbf67a4143c9c871412635a32d6249b3fb7b8f1e1ec87471c3abd88041078da3" exitCode=0 Dec 05 11:55:36 crc kubenswrapper[4809]: I1205 11:55:36.740260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cks67" event={"ID":"b2a801ec-eed1-4b07-bdb2-1654c69ffb98","Type":"ContainerDied","Data":"fbf67a4143c9c871412635a32d6249b3fb7b8f1e1ec87471c3abd88041078da3"} Dec 05 11:55:37 crc kubenswrapper[4809]: I1205 11:55:37.750296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cks67" event={"ID":"b2a801ec-eed1-4b07-bdb2-1654c69ffb98","Type":"ContainerStarted","Data":"6b44e6f037ddf1c91df336ca14d8af56e36f0fe5924c87ee6109f6e2864f1121"} Dec 05 11:55:37 crc kubenswrapper[4809]: I1205 11:55:37.770408 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cks67" podStartSLOduration=3.330647168 podStartE2EDuration="5.770384773s" podCreationTimestamp="2025-12-05 11:55:32 +0000 UTC" firstStartedPulling="2025-12-05 11:55:34.713144758 +0000 UTC m=+2830.104121316" lastFinishedPulling="2025-12-05 11:55:37.152882363 +0000 UTC m=+2832.543858921" observedRunningTime="2025-12-05 11:55:37.764898426 +0000 UTC m=+2833.155874974" watchObservedRunningTime="2025-12-05 11:55:37.770384773 +0000 UTC m=+2833.161361351" Dec 05 11:55:43 crc kubenswrapper[4809]: I1205 11:55:43.301272 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:43 crc kubenswrapper[4809]: I1205 11:55:43.301986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:43 crc kubenswrapper[4809]: I1205 11:55:43.364937 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:43 crc kubenswrapper[4809]: I1205 11:55:43.845149 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:43 crc kubenswrapper[4809]: I1205 11:55:43.899454 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cks67"] Dec 05 11:55:45 crc kubenswrapper[4809]: I1205 11:55:45.811728 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cks67" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="registry-server" containerID="cri-o://6b44e6f037ddf1c91df336ca14d8af56e36f0fe5924c87ee6109f6e2864f1121" gracePeriod=2 Dec 05 11:55:46 crc kubenswrapper[4809]: I1205 11:55:46.833396 4809 generic.go:334] "Generic (PLEG): container finished" podID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerID="6b44e6f037ddf1c91df336ca14d8af56e36f0fe5924c87ee6109f6e2864f1121" exitCode=0 Dec 05 11:55:46 crc kubenswrapper[4809]: I1205 11:55:46.833729 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cks67" event={"ID":"b2a801ec-eed1-4b07-bdb2-1654c69ffb98","Type":"ContainerDied","Data":"6b44e6f037ddf1c91df336ca14d8af56e36f0fe5924c87ee6109f6e2864f1121"} Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.322479 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.477733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gktrk\" (UniqueName: \"kubernetes.io/projected/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-kube-api-access-gktrk\") pod \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.477884 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-utilities\") pod \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.477952 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-catalog-content\") pod \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\" (UID: \"b2a801ec-eed1-4b07-bdb2-1654c69ffb98\") " Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.479420 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-utilities" (OuterVolumeSpecName: "utilities") pod "b2a801ec-eed1-4b07-bdb2-1654c69ffb98" (UID: "b2a801ec-eed1-4b07-bdb2-1654c69ffb98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.488012 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-kube-api-access-gktrk" (OuterVolumeSpecName: "kube-api-access-gktrk") pod "b2a801ec-eed1-4b07-bdb2-1654c69ffb98" (UID: "b2a801ec-eed1-4b07-bdb2-1654c69ffb98"). InnerVolumeSpecName "kube-api-access-gktrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.540037 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2a801ec-eed1-4b07-bdb2-1654c69ffb98" (UID: "b2a801ec-eed1-4b07-bdb2-1654c69ffb98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.579910 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gktrk\" (UniqueName: \"kubernetes.io/projected/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-kube-api-access-gktrk\") on node \"crc\" DevicePath \"\"" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.579963 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.579983 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a801ec-eed1-4b07-bdb2-1654c69ffb98-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.846363 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cks67" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.846365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cks67" event={"ID":"b2a801ec-eed1-4b07-bdb2-1654c69ffb98","Type":"ContainerDied","Data":"ba6f30eba66bf39d324086a5b40ec4384ae499dda9fed9321f1e5ffc934abbf1"} Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.846546 4809 scope.go:117] "RemoveContainer" containerID="6b44e6f037ddf1c91df336ca14d8af56e36f0fe5924c87ee6109f6e2864f1121" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.875586 4809 scope.go:117] "RemoveContainer" containerID="fbf67a4143c9c871412635a32d6249b3fb7b8f1e1ec87471c3abd88041078da3" Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.884267 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cks67"] Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.891478 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cks67"] Dec 05 11:55:47 crc kubenswrapper[4809]: I1205 11:55:47.918657 4809 scope.go:117] "RemoveContainer" containerID="737a7057f31c7eaec3ba1416f5ecf68ca0e721f79855b854efd8e6ab71d25d9e" Dec 05 11:55:48 crc kubenswrapper[4809]: I1205 11:55:48.885910 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" path="/var/lib/kubelet/pods/b2a801ec-eed1-4b07-bdb2-1654c69ffb98/volumes" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.830303 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bbtgj"] Dec 05 11:56:08 crc kubenswrapper[4809]: E1205 11:56:08.831333 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="extract-utilities" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.831348 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="extract-utilities" Dec 05 11:56:08 crc kubenswrapper[4809]: E1205 11:56:08.831367 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="registry-server" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.831375 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="registry-server" Dec 05 11:56:08 crc kubenswrapper[4809]: E1205 11:56:08.831387 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="extract-content" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.831395 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="extract-content" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.831599 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a801ec-eed1-4b07-bdb2-1654c69ffb98" containerName="registry-server" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.836355 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.843452 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bbtgj"] Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.998604 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-catalog-content\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.998868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-utilities\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:08 crc kubenswrapper[4809]: I1205 11:56:08.998942 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97pzk\" (UniqueName: \"kubernetes.io/projected/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-kube-api-access-97pzk\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.100085 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-catalog-content\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.100176 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-utilities\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.100207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97pzk\" (UniqueName: \"kubernetes.io/projected/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-kube-api-access-97pzk\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.100731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-catalog-content\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.100760 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-utilities\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.119457 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97pzk\" (UniqueName: \"kubernetes.io/projected/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-kube-api-access-97pzk\") pod \"certified-operators-bbtgj\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.158344 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:09 crc kubenswrapper[4809]: I1205 11:56:09.653060 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bbtgj"] Dec 05 11:56:10 crc kubenswrapper[4809]: I1205 11:56:10.017980 4809 generic.go:334] "Generic (PLEG): container finished" podID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerID="d2c12c20b6bc96ea8b96fa8b6bc80efeb9b56eea0120a58704f832cc267cf817" exitCode=0 Dec 05 11:56:10 crc kubenswrapper[4809]: I1205 11:56:10.018022 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbtgj" event={"ID":"15dad7ce-6ee7-466c-92da-8e4e4e5beccf","Type":"ContainerDied","Data":"d2c12c20b6bc96ea8b96fa8b6bc80efeb9b56eea0120a58704f832cc267cf817"} Dec 05 11:56:10 crc kubenswrapper[4809]: I1205 11:56:10.019715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbtgj" event={"ID":"15dad7ce-6ee7-466c-92da-8e4e4e5beccf","Type":"ContainerStarted","Data":"871325aede3ce49729efd4ef88685d639d9b76dbf07de8298262446595e44897"} Dec 05 11:56:11 crc kubenswrapper[4809]: I1205 11:56:11.032720 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbtgj" event={"ID":"15dad7ce-6ee7-466c-92da-8e4e4e5beccf","Type":"ContainerStarted","Data":"daab3ba9fe7be23a954a61dc3b43a9dce054626654e7c1f8715fe4e1415e2853"} Dec 05 11:56:12 crc kubenswrapper[4809]: I1205 11:56:12.047834 4809 generic.go:334] "Generic (PLEG): container finished" podID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerID="daab3ba9fe7be23a954a61dc3b43a9dce054626654e7c1f8715fe4e1415e2853" exitCode=0 Dec 05 11:56:12 crc kubenswrapper[4809]: I1205 11:56:12.048174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbtgj" event={"ID":"15dad7ce-6ee7-466c-92da-8e4e4e5beccf","Type":"ContainerDied","Data":"daab3ba9fe7be23a954a61dc3b43a9dce054626654e7c1f8715fe4e1415e2853"} Dec 05 11:56:13 crc kubenswrapper[4809]: I1205 11:56:13.057432 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbtgj" event={"ID":"15dad7ce-6ee7-466c-92da-8e4e4e5beccf","Type":"ContainerStarted","Data":"e20db48dd39ce9ee6bffe50e3bd7fc7a4ef9b095261cdebadd383b54f2d588cd"} Dec 05 11:56:13 crc kubenswrapper[4809]: I1205 11:56:13.077570 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bbtgj" podStartSLOduration=2.530671049 podStartE2EDuration="5.077554038s" podCreationTimestamp="2025-12-05 11:56:08 +0000 UTC" firstStartedPulling="2025-12-05 11:56:10.019114574 +0000 UTC m=+2865.410091132" lastFinishedPulling="2025-12-05 11:56:12.565997563 +0000 UTC m=+2867.956974121" observedRunningTime="2025-12-05 11:56:13.074246779 +0000 UTC m=+2868.465223327" watchObservedRunningTime="2025-12-05 11:56:13.077554038 +0000 UTC m=+2868.468530596" Dec 05 11:56:14 crc kubenswrapper[4809]: I1205 11:56:14.046703 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:56:14 crc kubenswrapper[4809]: I1205 11:56:14.046830 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:56:19 crc kubenswrapper[4809]: I1205 11:56:19.159143 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:19 crc kubenswrapper[4809]: I1205 11:56:19.159485 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:19 crc kubenswrapper[4809]: I1205 11:56:19.201797 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:20 crc kubenswrapper[4809]: I1205 11:56:20.147078 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:20 crc kubenswrapper[4809]: I1205 11:56:20.199194 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bbtgj"] Dec 05 11:56:22 crc kubenswrapper[4809]: I1205 11:56:22.118468 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bbtgj" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="registry-server" containerID="cri-o://e20db48dd39ce9ee6bffe50e3bd7fc7a4ef9b095261cdebadd383b54f2d588cd" gracePeriod=2 Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.137074 4809 generic.go:334] "Generic (PLEG): container finished" podID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerID="e20db48dd39ce9ee6bffe50e3bd7fc7a4ef9b095261cdebadd383b54f2d588cd" exitCode=0 Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.137124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbtgj" event={"ID":"15dad7ce-6ee7-466c-92da-8e4e4e5beccf","Type":"ContainerDied","Data":"e20db48dd39ce9ee6bffe50e3bd7fc7a4ef9b095261cdebadd383b54f2d588cd"} Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.384555 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.541463 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97pzk\" (UniqueName: \"kubernetes.io/projected/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-kube-api-access-97pzk\") pod \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.541518 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-catalog-content\") pod \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.541542 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-utilities\") pod \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\" (UID: \"15dad7ce-6ee7-466c-92da-8e4e4e5beccf\") " Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.542733 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-utilities" (OuterVolumeSpecName: "utilities") pod "15dad7ce-6ee7-466c-92da-8e4e4e5beccf" (UID: "15dad7ce-6ee7-466c-92da-8e4e4e5beccf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.547146 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-kube-api-access-97pzk" (OuterVolumeSpecName: "kube-api-access-97pzk") pod "15dad7ce-6ee7-466c-92da-8e4e4e5beccf" (UID: "15dad7ce-6ee7-466c-92da-8e4e4e5beccf"). InnerVolumeSpecName "kube-api-access-97pzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.594685 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15dad7ce-6ee7-466c-92da-8e4e4e5beccf" (UID: "15dad7ce-6ee7-466c-92da-8e4e4e5beccf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.642447 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97pzk\" (UniqueName: \"kubernetes.io/projected/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-kube-api-access-97pzk\") on node \"crc\" DevicePath \"\"" Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.642477 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:56:24 crc kubenswrapper[4809]: I1205 11:56:24.642486 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dad7ce-6ee7-466c-92da-8e4e4e5beccf-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:56:25 crc kubenswrapper[4809]: I1205 11:56:25.148148 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bbtgj" event={"ID":"15dad7ce-6ee7-466c-92da-8e4e4e5beccf","Type":"ContainerDied","Data":"871325aede3ce49729efd4ef88685d639d9b76dbf07de8298262446595e44897"} Dec 05 11:56:25 crc kubenswrapper[4809]: I1205 11:56:25.148476 4809 scope.go:117] "RemoveContainer" containerID="e20db48dd39ce9ee6bffe50e3bd7fc7a4ef9b095261cdebadd383b54f2d588cd" Dec 05 11:56:25 crc kubenswrapper[4809]: I1205 11:56:25.148294 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bbtgj" Dec 05 11:56:25 crc kubenswrapper[4809]: I1205 11:56:25.174400 4809 scope.go:117] "RemoveContainer" containerID="daab3ba9fe7be23a954a61dc3b43a9dce054626654e7c1f8715fe4e1415e2853" Dec 05 11:56:25 crc kubenswrapper[4809]: I1205 11:56:25.178616 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bbtgj"] Dec 05 11:56:25 crc kubenswrapper[4809]: I1205 11:56:25.187229 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bbtgj"] Dec 05 11:56:25 crc kubenswrapper[4809]: I1205 11:56:25.194778 4809 scope.go:117] "RemoveContainer" containerID="d2c12c20b6bc96ea8b96fa8b6bc80efeb9b56eea0120a58704f832cc267cf817" Dec 05 11:56:26 crc kubenswrapper[4809]: I1205 11:56:26.887243 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" path="/var/lib/kubelet/pods/15dad7ce-6ee7-466c-92da-8e4e4e5beccf/volumes" Dec 05 11:56:44 crc kubenswrapper[4809]: I1205 11:56:44.046400 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:56:44 crc kubenswrapper[4809]: I1205 11:56:44.048320 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.080559 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-khsgr"] Dec 05 11:56:53 crc kubenswrapper[4809]: E1205 11:56:53.081935 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="registry-server" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.081958 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="registry-server" Dec 05 11:56:53 crc kubenswrapper[4809]: E1205 11:56:53.081977 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="extract-content" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.081989 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="extract-content" Dec 05 11:56:53 crc kubenswrapper[4809]: E1205 11:56:53.082046 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="extract-utilities" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.082059 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="extract-utilities" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.082344 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="15dad7ce-6ee7-466c-92da-8e4e4e5beccf" containerName="registry-server" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.084274 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.088621 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-khsgr"] Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.202339 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-utilities\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.202543 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-catalog-content\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.202764 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl5br\" (UniqueName: \"kubernetes.io/projected/ff6a187e-40a4-4109-aec6-83e39b270186-kube-api-access-pl5br\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.303857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-utilities\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.303925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-catalog-content\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.303979 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl5br\" (UniqueName: \"kubernetes.io/projected/ff6a187e-40a4-4109-aec6-83e39b270186-kube-api-access-pl5br\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.304382 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-utilities\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.304488 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-catalog-content\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.326503 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl5br\" (UniqueName: \"kubernetes.io/projected/ff6a187e-40a4-4109-aec6-83e39b270186-kube-api-access-pl5br\") pod \"redhat-operators-khsgr\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.424270 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:56:53 crc kubenswrapper[4809]: I1205 11:56:53.878586 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-khsgr"] Dec 05 11:56:54 crc kubenswrapper[4809]: I1205 11:56:54.385451 4809 generic.go:334] "Generic (PLEG): container finished" podID="ff6a187e-40a4-4109-aec6-83e39b270186" containerID="bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6" exitCode=0 Dec 05 11:56:54 crc kubenswrapper[4809]: I1205 11:56:54.385507 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khsgr" event={"ID":"ff6a187e-40a4-4109-aec6-83e39b270186","Type":"ContainerDied","Data":"bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6"} Dec 05 11:56:54 crc kubenswrapper[4809]: I1205 11:56:54.385536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khsgr" event={"ID":"ff6a187e-40a4-4109-aec6-83e39b270186","Type":"ContainerStarted","Data":"98341472e5e2680bb7b9b55e80146c960b6bdba2851cc4ec81d1ea48e28d2ea7"} Dec 05 11:56:55 crc kubenswrapper[4809]: I1205 11:56:55.398426 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khsgr" event={"ID":"ff6a187e-40a4-4109-aec6-83e39b270186","Type":"ContainerStarted","Data":"31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f"} Dec 05 11:56:56 crc kubenswrapper[4809]: I1205 11:56:56.410821 4809 generic.go:334] "Generic (PLEG): container finished" podID="ff6a187e-40a4-4109-aec6-83e39b270186" containerID="31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f" exitCode=0 Dec 05 11:56:56 crc kubenswrapper[4809]: I1205 11:56:56.410883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khsgr" event={"ID":"ff6a187e-40a4-4109-aec6-83e39b270186","Type":"ContainerDied","Data":"31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f"} Dec 05 11:56:57 crc kubenswrapper[4809]: I1205 11:56:57.422931 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khsgr" event={"ID":"ff6a187e-40a4-4109-aec6-83e39b270186","Type":"ContainerStarted","Data":"e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3"} Dec 05 11:56:57 crc kubenswrapper[4809]: I1205 11:56:57.463497 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-khsgr" podStartSLOduration=2.070736005 podStartE2EDuration="4.463469598s" podCreationTimestamp="2025-12-05 11:56:53 +0000 UTC" firstStartedPulling="2025-12-05 11:56:54.38694397 +0000 UTC m=+2909.777920528" lastFinishedPulling="2025-12-05 11:56:56.779677563 +0000 UTC m=+2912.170654121" observedRunningTime="2025-12-05 11:56:57.451390615 +0000 UTC m=+2912.842367213" watchObservedRunningTime="2025-12-05 11:56:57.463469598 +0000 UTC m=+2912.854446196" Dec 05 11:57:03 crc kubenswrapper[4809]: I1205 11:57:03.425277 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:57:03 crc kubenswrapper[4809]: I1205 11:57:03.425886 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:57:03 crc kubenswrapper[4809]: I1205 11:57:03.474404 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:57:03 crc kubenswrapper[4809]: I1205 11:57:03.524657 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:57:05 crc kubenswrapper[4809]: I1205 11:57:05.945608 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-khsgr"] Dec 05 11:57:05 crc kubenswrapper[4809]: I1205 11:57:05.946235 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-khsgr" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="registry-server" containerID="cri-o://e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3" gracePeriod=2 Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.070141 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.250065 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-catalog-content\") pod \"ff6a187e-40a4-4109-aec6-83e39b270186\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.250182 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl5br\" (UniqueName: \"kubernetes.io/projected/ff6a187e-40a4-4109-aec6-83e39b270186-kube-api-access-pl5br\") pod \"ff6a187e-40a4-4109-aec6-83e39b270186\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.250367 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-utilities\") pod \"ff6a187e-40a4-4109-aec6-83e39b270186\" (UID: \"ff6a187e-40a4-4109-aec6-83e39b270186\") " Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.251552 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-utilities" (OuterVolumeSpecName: "utilities") pod "ff6a187e-40a4-4109-aec6-83e39b270186" (UID: "ff6a187e-40a4-4109-aec6-83e39b270186"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.257784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff6a187e-40a4-4109-aec6-83e39b270186-kube-api-access-pl5br" (OuterVolumeSpecName: "kube-api-access-pl5br") pod "ff6a187e-40a4-4109-aec6-83e39b270186" (UID: "ff6a187e-40a4-4109-aec6-83e39b270186"). InnerVolumeSpecName "kube-api-access-pl5br". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.351800 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.351833 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl5br\" (UniqueName: \"kubernetes.io/projected/ff6a187e-40a4-4109-aec6-83e39b270186-kube-api-access-pl5br\") on node \"crc\" DevicePath \"\"" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.380314 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff6a187e-40a4-4109-aec6-83e39b270186" (UID: "ff6a187e-40a4-4109-aec6-83e39b270186"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.453765 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6a187e-40a4-4109-aec6-83e39b270186-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.527855 4809 generic.go:334] "Generic (PLEG): container finished" podID="ff6a187e-40a4-4109-aec6-83e39b270186" containerID="e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3" exitCode=0 Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.527899 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khsgr" event={"ID":"ff6a187e-40a4-4109-aec6-83e39b270186","Type":"ContainerDied","Data":"e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3"} Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.527930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khsgr" event={"ID":"ff6a187e-40a4-4109-aec6-83e39b270186","Type":"ContainerDied","Data":"98341472e5e2680bb7b9b55e80146c960b6bdba2851cc4ec81d1ea48e28d2ea7"} Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.527953 4809 scope.go:117] "RemoveContainer" containerID="e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.528096 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-khsgr" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.564431 4809 scope.go:117] "RemoveContainer" containerID="31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.582136 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-khsgr"] Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.588028 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-khsgr"] Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.591855 4809 scope.go:117] "RemoveContainer" containerID="bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.617431 4809 scope.go:117] "RemoveContainer" containerID="e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3" Dec 05 11:57:09 crc kubenswrapper[4809]: E1205 11:57:09.617844 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3\": container with ID starting with e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3 not found: ID does not exist" containerID="e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.617878 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3"} err="failed to get container status \"e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3\": rpc error: code = NotFound desc = could not find container \"e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3\": container with ID starting with e044c8a03b8a9bd83d3c66465376cc7f02172e69bf9b1b0ab4625b31ede07cf3 not found: ID does not exist" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.617905 4809 scope.go:117] "RemoveContainer" containerID="31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f" Dec 05 11:57:09 crc kubenswrapper[4809]: E1205 11:57:09.618624 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f\": container with ID starting with 31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f not found: ID does not exist" containerID="31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.618662 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f"} err="failed to get container status \"31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f\": rpc error: code = NotFound desc = could not find container \"31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f\": container with ID starting with 31e189ccebdfd424ce482917272f3d7de82da00cbd6c623fa4bc3efbd036246f not found: ID does not exist" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.618677 4809 scope.go:117] "RemoveContainer" containerID="bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6" Dec 05 11:57:09 crc kubenswrapper[4809]: E1205 11:57:09.618911 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6\": container with ID starting with bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6 not found: ID does not exist" containerID="bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6" Dec 05 11:57:09 crc kubenswrapper[4809]: I1205 11:57:09.618934 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6"} err="failed to get container status \"bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6\": rpc error: code = NotFound desc = could not find container \"bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6\": container with ID starting with bb4430c1bb035e7a00d48bbaa6927a8f457c27463a9f6dc25b13acd6e62a1cb6 not found: ID does not exist" Dec 05 11:57:10 crc kubenswrapper[4809]: I1205 11:57:10.880691 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" path="/var/lib/kubelet/pods/ff6a187e-40a4-4109-aec6-83e39b270186/volumes" Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.046936 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.047335 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.047388 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.048057 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.048131 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" gracePeriod=600 Dec 05 11:57:14 crc kubenswrapper[4809]: E1205 11:57:14.173241 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.574411 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" exitCode=0 Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.574494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da"} Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.574925 4809 scope.go:117] "RemoveContainer" containerID="6da7f8756648d6aa0d5dcf32827c2417d82a2103f968e117591720c658b834b2" Dec 05 11:57:14 crc kubenswrapper[4809]: I1205 11:57:14.575602 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:57:14 crc kubenswrapper[4809]: E1205 11:57:14.576308 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:57:28 crc kubenswrapper[4809]: I1205 11:57:28.872645 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:57:28 crc kubenswrapper[4809]: E1205 11:57:28.873451 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:57:40 crc kubenswrapper[4809]: I1205 11:57:40.872519 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:57:40 crc kubenswrapper[4809]: E1205 11:57:40.873366 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:57:54 crc kubenswrapper[4809]: I1205 11:57:54.880297 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:57:54 crc kubenswrapper[4809]: E1205 11:57:54.881191 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:58:08 crc kubenswrapper[4809]: I1205 11:58:08.873389 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:58:08 crc kubenswrapper[4809]: E1205 11:58:08.874515 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:58:21 crc kubenswrapper[4809]: I1205 11:58:21.871889 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:58:21 crc kubenswrapper[4809]: E1205 11:58:21.873158 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:58:36 crc kubenswrapper[4809]: I1205 11:58:36.873344 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:58:36 crc kubenswrapper[4809]: E1205 11:58:36.874142 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:58:47 crc kubenswrapper[4809]: I1205 11:58:47.872702 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:58:47 crc kubenswrapper[4809]: E1205 11:58:47.873546 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:59:01 crc kubenswrapper[4809]: I1205 11:59:01.872086 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:59:01 crc kubenswrapper[4809]: E1205 11:59:01.872857 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:59:14 crc kubenswrapper[4809]: I1205 11:59:14.876832 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:59:14 crc kubenswrapper[4809]: E1205 11:59:14.878264 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:59:27 crc kubenswrapper[4809]: I1205 11:59:27.873412 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:59:27 crc kubenswrapper[4809]: E1205 11:59:27.874700 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:59:39 crc kubenswrapper[4809]: I1205 11:59:39.871945 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:59:39 crc kubenswrapper[4809]: E1205 11:59:39.873016 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 11:59:53 crc kubenswrapper[4809]: I1205 11:59:53.872575 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 11:59:53 crc kubenswrapper[4809]: E1205 11:59:53.873627 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.155854 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr"] Dec 05 12:00:00 crc kubenswrapper[4809]: E1205 12:00:00.157130 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="extract-content" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.157160 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="extract-content" Dec 05 12:00:00 crc kubenswrapper[4809]: E1205 12:00:00.172948 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="extract-utilities" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.172978 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="extract-utilities" Dec 05 12:00:00 crc kubenswrapper[4809]: E1205 12:00:00.172998 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="registry-server" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.173007 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="registry-server" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.174258 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff6a187e-40a4-4109-aec6-83e39b270186" containerName="registry-server" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.175314 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr"] Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.175480 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.187846 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.188563 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.237414 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/daa47f26-d043-470e-a6d9-030daba81545-secret-volume\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.237565 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/daa47f26-d043-470e-a6d9-030daba81545-config-volume\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.237659 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dpcr\" (UniqueName: \"kubernetes.io/projected/daa47f26-d043-470e-a6d9-030daba81545-kube-api-access-7dpcr\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.339081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/daa47f26-d043-470e-a6d9-030daba81545-secret-volume\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.339550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/daa47f26-d043-470e-a6d9-030daba81545-config-volume\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.339614 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dpcr\" (UniqueName: \"kubernetes.io/projected/daa47f26-d043-470e-a6d9-030daba81545-kube-api-access-7dpcr\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.340762 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/daa47f26-d043-470e-a6d9-030daba81545-config-volume\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.348977 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/daa47f26-d043-470e-a6d9-030daba81545-secret-volume\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.357864 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dpcr\" (UniqueName: \"kubernetes.io/projected/daa47f26-d043-470e-a6d9-030daba81545-kube-api-access-7dpcr\") pod \"collect-profiles-29415600-kv5lr\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.508723 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:00 crc kubenswrapper[4809]: I1205 12:00:00.967439 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr"] Dec 05 12:00:01 crc kubenswrapper[4809]: I1205 12:00:01.934769 4809 generic.go:334] "Generic (PLEG): container finished" podID="daa47f26-d043-470e-a6d9-030daba81545" containerID="fe39dec510a18d985b99fdac8a723904afc0b00035aa86aa9c60c37824a06055" exitCode=0 Dec 05 12:00:01 crc kubenswrapper[4809]: I1205 12:00:01.934852 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" event={"ID":"daa47f26-d043-470e-a6d9-030daba81545","Type":"ContainerDied","Data":"fe39dec510a18d985b99fdac8a723904afc0b00035aa86aa9c60c37824a06055"} Dec 05 12:00:01 crc kubenswrapper[4809]: I1205 12:00:01.935125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" event={"ID":"daa47f26-d043-470e-a6d9-030daba81545","Type":"ContainerStarted","Data":"b403b0518e7740e1b6fbe9615b37fc57181d00fbf6a350d7d5b36a6824bff735"} Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.233901 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.388685 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/daa47f26-d043-470e-a6d9-030daba81545-secret-volume\") pod \"daa47f26-d043-470e-a6d9-030daba81545\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.388782 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/daa47f26-d043-470e-a6d9-030daba81545-config-volume\") pod \"daa47f26-d043-470e-a6d9-030daba81545\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.388831 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dpcr\" (UniqueName: \"kubernetes.io/projected/daa47f26-d043-470e-a6d9-030daba81545-kube-api-access-7dpcr\") pod \"daa47f26-d043-470e-a6d9-030daba81545\" (UID: \"daa47f26-d043-470e-a6d9-030daba81545\") " Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.389591 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daa47f26-d043-470e-a6d9-030daba81545-config-volume" (OuterVolumeSpecName: "config-volume") pod "daa47f26-d043-470e-a6d9-030daba81545" (UID: "daa47f26-d043-470e-a6d9-030daba81545"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.394113 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daa47f26-d043-470e-a6d9-030daba81545-kube-api-access-7dpcr" (OuterVolumeSpecName: "kube-api-access-7dpcr") pod "daa47f26-d043-470e-a6d9-030daba81545" (UID: "daa47f26-d043-470e-a6d9-030daba81545"). InnerVolumeSpecName "kube-api-access-7dpcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.394890 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daa47f26-d043-470e-a6d9-030daba81545-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "daa47f26-d043-470e-a6d9-030daba81545" (UID: "daa47f26-d043-470e-a6d9-030daba81545"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.490866 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/daa47f26-d043-470e-a6d9-030daba81545-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.490903 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dpcr\" (UniqueName: \"kubernetes.io/projected/daa47f26-d043-470e-a6d9-030daba81545-kube-api-access-7dpcr\") on node \"crc\" DevicePath \"\"" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.490914 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/daa47f26-d043-470e-a6d9-030daba81545-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.950214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" event={"ID":"daa47f26-d043-470e-a6d9-030daba81545","Type":"ContainerDied","Data":"b403b0518e7740e1b6fbe9615b37fc57181d00fbf6a350d7d5b36a6824bff735"} Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.950264 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr" Dec 05 12:00:03 crc kubenswrapper[4809]: I1205 12:00:03.950288 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b403b0518e7740e1b6fbe9615b37fc57181d00fbf6a350d7d5b36a6824bff735" Dec 05 12:00:04 crc kubenswrapper[4809]: I1205 12:00:04.309678 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg"] Dec 05 12:00:04 crc kubenswrapper[4809]: I1205 12:00:04.317556 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-jcsgg"] Dec 05 12:00:04 crc kubenswrapper[4809]: I1205 12:00:04.877280 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:00:04 crc kubenswrapper[4809]: E1205 12:00:04.877808 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:00:04 crc kubenswrapper[4809]: I1205 12:00:04.882887 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56f75595-2600-4f31-9204-c94f4d6e212f" path="/var/lib/kubelet/pods/56f75595-2600-4f31-9204-c94f4d6e212f/volumes" Dec 05 12:00:16 crc kubenswrapper[4809]: I1205 12:00:16.872053 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:00:16 crc kubenswrapper[4809]: E1205 12:00:16.872884 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:00:28 crc kubenswrapper[4809]: I1205 12:00:28.872273 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:00:28 crc kubenswrapper[4809]: E1205 12:00:28.873006 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:00:33 crc kubenswrapper[4809]: I1205 12:00:33.385853 4809 scope.go:117] "RemoveContainer" containerID="b700cfa536a2b49e4d9c37807eab2f08d31a79547d9f084f54e09ec0ed45057f" Dec 05 12:00:42 crc kubenswrapper[4809]: I1205 12:00:42.876355 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:00:42 crc kubenswrapper[4809]: E1205 12:00:42.877257 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:00:55 crc kubenswrapper[4809]: I1205 12:00:55.871910 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:00:55 crc kubenswrapper[4809]: E1205 12:00:55.872744 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:01:10 crc kubenswrapper[4809]: I1205 12:01:10.871971 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:01:10 crc kubenswrapper[4809]: E1205 12:01:10.872817 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:01:23 crc kubenswrapper[4809]: I1205 12:01:23.872331 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:01:23 crc kubenswrapper[4809]: E1205 12:01:23.873442 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:01:35 crc kubenswrapper[4809]: I1205 12:01:35.874018 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:01:35 crc kubenswrapper[4809]: E1205 12:01:35.874820 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:01:46 crc kubenswrapper[4809]: I1205 12:01:46.872748 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:01:46 crc kubenswrapper[4809]: E1205 12:01:46.873486 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:02:00 crc kubenswrapper[4809]: I1205 12:02:00.872750 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:02:00 crc kubenswrapper[4809]: E1205 12:02:00.873874 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:02:14 crc kubenswrapper[4809]: I1205 12:02:14.878233 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:02:15 crc kubenswrapper[4809]: I1205 12:02:15.943512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"d61be0cb10852d25dee858d34fd89106d48f1e911a9ce1010259ac7fc3a322f9"} Dec 05 12:04:44 crc kubenswrapper[4809]: I1205 12:04:44.046748 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:04:44 crc kubenswrapper[4809]: I1205 12:04:44.047438 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:05:14 crc kubenswrapper[4809]: I1205 12:05:14.046479 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:05:14 crc kubenswrapper[4809]: I1205 12:05:14.047031 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.046825 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.047400 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.047450 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.048131 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d61be0cb10852d25dee858d34fd89106d48f1e911a9ce1010259ac7fc3a322f9"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.048188 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://d61be0cb10852d25dee858d34fd89106d48f1e911a9ce1010259ac7fc3a322f9" gracePeriod=600 Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.186706 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="d61be0cb10852d25dee858d34fd89106d48f1e911a9ce1010259ac7fc3a322f9" exitCode=0 Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.186854 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"d61be0cb10852d25dee858d34fd89106d48f1e911a9ce1010259ac7fc3a322f9"} Dec 05 12:05:44 crc kubenswrapper[4809]: I1205 12:05:44.187243 4809 scope.go:117] "RemoveContainer" containerID="25b06f4edcd626ebf1fd093b4403e44a47e4d56f006ae31e33ebf7c9050194da" Dec 05 12:05:45 crc kubenswrapper[4809]: I1205 12:05:45.195982 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2"} Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.604827 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rxmv7"] Dec 05 12:06:07 crc kubenswrapper[4809]: E1205 12:06:07.605817 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa47f26-d043-470e-a6d9-030daba81545" containerName="collect-profiles" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.605835 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa47f26-d043-470e-a6d9-030daba81545" containerName="collect-profiles" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.605982 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="daa47f26-d043-470e-a6d9-030daba81545" containerName="collect-profiles" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.607273 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.633974 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxmv7"] Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.639372 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-catalog-content\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.639416 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbqvg\" (UniqueName: \"kubernetes.io/projected/29fd3dc9-7d75-411e-bc60-8843cab48a33-kube-api-access-bbqvg\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.639508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-utilities\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.740181 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-utilities\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.740244 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-catalog-content\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.740263 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbqvg\" (UniqueName: \"kubernetes.io/projected/29fd3dc9-7d75-411e-bc60-8843cab48a33-kube-api-access-bbqvg\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.741009 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-catalog-content\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.741067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-utilities\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.762982 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbqvg\" (UniqueName: \"kubernetes.io/projected/29fd3dc9-7d75-411e-bc60-8843cab48a33-kube-api-access-bbqvg\") pod \"redhat-marketplace-rxmv7\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:07 crc kubenswrapper[4809]: I1205 12:06:07.939517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:08 crc kubenswrapper[4809]: I1205 12:06:08.211125 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxmv7"] Dec 05 12:06:08 crc kubenswrapper[4809]: I1205 12:06:08.387755 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxmv7" event={"ID":"29fd3dc9-7d75-411e-bc60-8843cab48a33","Type":"ContainerStarted","Data":"baef4d7c2346d2bef9313020037abde01c6a1cc7ee84f9a357aee4964a7b41a4"} Dec 05 12:06:09 crc kubenswrapper[4809]: I1205 12:06:09.397450 4809 generic.go:334] "Generic (PLEG): container finished" podID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerID="1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85" exitCode=0 Dec 05 12:06:09 crc kubenswrapper[4809]: I1205 12:06:09.397507 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxmv7" event={"ID":"29fd3dc9-7d75-411e-bc60-8843cab48a33","Type":"ContainerDied","Data":"1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85"} Dec 05 12:06:09 crc kubenswrapper[4809]: I1205 12:06:09.401143 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:06:10 crc kubenswrapper[4809]: I1205 12:06:10.408386 4809 generic.go:334] "Generic (PLEG): container finished" podID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerID="42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3" exitCode=0 Dec 05 12:06:10 crc kubenswrapper[4809]: I1205 12:06:10.408475 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxmv7" event={"ID":"29fd3dc9-7d75-411e-bc60-8843cab48a33","Type":"ContainerDied","Data":"42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3"} Dec 05 12:06:11 crc kubenswrapper[4809]: I1205 12:06:11.420179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxmv7" event={"ID":"29fd3dc9-7d75-411e-bc60-8843cab48a33","Type":"ContainerStarted","Data":"1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035"} Dec 05 12:06:11 crc kubenswrapper[4809]: I1205 12:06:11.441856 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rxmv7" podStartSLOduration=3.041880599 podStartE2EDuration="4.441837572s" podCreationTimestamp="2025-12-05 12:06:07 +0000 UTC" firstStartedPulling="2025-12-05 12:06:09.400920455 +0000 UTC m=+3464.791897013" lastFinishedPulling="2025-12-05 12:06:10.800877438 +0000 UTC m=+3466.191853986" observedRunningTime="2025-12-05 12:06:11.43767014 +0000 UTC m=+3466.828646718" watchObservedRunningTime="2025-12-05 12:06:11.441837572 +0000 UTC m=+3466.832814130" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.580247 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zwjfw"] Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.582861 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.611997 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwjfw"] Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.730223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d699\" (UniqueName: \"kubernetes.io/projected/ae8536c1-4129-415f-aabc-7b8bff16962d-kube-api-access-5d699\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.730410 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-utilities\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.730494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-catalog-content\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.832141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-utilities\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.832212 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-catalog-content\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.832262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d699\" (UniqueName: \"kubernetes.io/projected/ae8536c1-4129-415f-aabc-7b8bff16962d-kube-api-access-5d699\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.832717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-utilities\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.832914 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-catalog-content\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.854098 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d699\" (UniqueName: \"kubernetes.io/projected/ae8536c1-4129-415f-aabc-7b8bff16962d-kube-api-access-5d699\") pod \"community-operators-zwjfw\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:13 crc kubenswrapper[4809]: I1205 12:06:13.907930 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:14 crc kubenswrapper[4809]: I1205 12:06:14.444269 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwjfw"] Dec 05 12:06:15 crc kubenswrapper[4809]: I1205 12:06:15.469515 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerID="62f824397585fbb9c780b814d38c8125c909e91fd09aea9b1ef3e350af8c1779" exitCode=0 Dec 05 12:06:15 crc kubenswrapper[4809]: I1205 12:06:15.469628 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjfw" event={"ID":"ae8536c1-4129-415f-aabc-7b8bff16962d","Type":"ContainerDied","Data":"62f824397585fbb9c780b814d38c8125c909e91fd09aea9b1ef3e350af8c1779"} Dec 05 12:06:15 crc kubenswrapper[4809]: I1205 12:06:15.469814 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjfw" event={"ID":"ae8536c1-4129-415f-aabc-7b8bff16962d","Type":"ContainerStarted","Data":"0e04a8c844ab1929c9b72458602a87640d9f2a7c18c90433710e6f73ffe55d8f"} Dec 05 12:06:17 crc kubenswrapper[4809]: I1205 12:06:17.485359 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerID="6aca686601170362635755f9585e77ff277e2322b6e15c76ec9751c5fdee2550" exitCode=0 Dec 05 12:06:17 crc kubenswrapper[4809]: I1205 12:06:17.485568 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjfw" event={"ID":"ae8536c1-4129-415f-aabc-7b8bff16962d","Type":"ContainerDied","Data":"6aca686601170362635755f9585e77ff277e2322b6e15c76ec9751c5fdee2550"} Dec 05 12:06:17 crc kubenswrapper[4809]: I1205 12:06:17.939939 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:17 crc kubenswrapper[4809]: I1205 12:06:17.940985 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:18 crc kubenswrapper[4809]: I1205 12:06:18.010298 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:18 crc kubenswrapper[4809]: I1205 12:06:18.497244 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjfw" event={"ID":"ae8536c1-4129-415f-aabc-7b8bff16962d","Type":"ContainerStarted","Data":"18b6f9f42d4546b892ad8b53a7bd41a67543dc3e054572bbbdba2eea8e1a2bfd"} Dec 05 12:06:18 crc kubenswrapper[4809]: I1205 12:06:18.529392 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zwjfw" podStartSLOduration=3.00892315 podStartE2EDuration="5.529359667s" podCreationTimestamp="2025-12-05 12:06:13 +0000 UTC" firstStartedPulling="2025-12-05 12:06:15.472047896 +0000 UTC m=+3470.863024454" lastFinishedPulling="2025-12-05 12:06:17.992484413 +0000 UTC m=+3473.383460971" observedRunningTime="2025-12-05 12:06:18.516808599 +0000 UTC m=+3473.907785157" watchObservedRunningTime="2025-12-05 12:06:18.529359667 +0000 UTC m=+3473.920336225" Dec 05 12:06:18 crc kubenswrapper[4809]: I1205 12:06:18.553239 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:19 crc kubenswrapper[4809]: I1205 12:06:19.772530 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxmv7"] Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.516358 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rxmv7" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="registry-server" containerID="cri-o://1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035" gracePeriod=2 Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.915401 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.963856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbqvg\" (UniqueName: \"kubernetes.io/projected/29fd3dc9-7d75-411e-bc60-8843cab48a33-kube-api-access-bbqvg\") pod \"29fd3dc9-7d75-411e-bc60-8843cab48a33\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.964048 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-utilities\") pod \"29fd3dc9-7d75-411e-bc60-8843cab48a33\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.964164 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-catalog-content\") pod \"29fd3dc9-7d75-411e-bc60-8843cab48a33\" (UID: \"29fd3dc9-7d75-411e-bc60-8843cab48a33\") " Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.966295 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-utilities" (OuterVolumeSpecName: "utilities") pod "29fd3dc9-7d75-411e-bc60-8843cab48a33" (UID: "29fd3dc9-7d75-411e-bc60-8843cab48a33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.972657 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29fd3dc9-7d75-411e-bc60-8843cab48a33-kube-api-access-bbqvg" (OuterVolumeSpecName: "kube-api-access-bbqvg") pod "29fd3dc9-7d75-411e-bc60-8843cab48a33" (UID: "29fd3dc9-7d75-411e-bc60-8843cab48a33"). InnerVolumeSpecName "kube-api-access-bbqvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:21 crc kubenswrapper[4809]: I1205 12:06:21.989076 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29fd3dc9-7d75-411e-bc60-8843cab48a33" (UID: "29fd3dc9-7d75-411e-bc60-8843cab48a33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.066414 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbqvg\" (UniqueName: \"kubernetes.io/projected/29fd3dc9-7d75-411e-bc60-8843cab48a33-kube-api-access-bbqvg\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.066508 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.066525 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fd3dc9-7d75-411e-bc60-8843cab48a33-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.525561 4809 generic.go:334] "Generic (PLEG): container finished" podID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerID="1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035" exitCode=0 Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.525601 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxmv7" event={"ID":"29fd3dc9-7d75-411e-bc60-8843cab48a33","Type":"ContainerDied","Data":"1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035"} Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.525622 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxmv7" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.525667 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxmv7" event={"ID":"29fd3dc9-7d75-411e-bc60-8843cab48a33","Type":"ContainerDied","Data":"baef4d7c2346d2bef9313020037abde01c6a1cc7ee84f9a357aee4964a7b41a4"} Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.525690 4809 scope.go:117] "RemoveContainer" containerID="1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.560079 4809 scope.go:117] "RemoveContainer" containerID="42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.560875 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxmv7"] Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.577306 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxmv7"] Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.581817 4809 scope.go:117] "RemoveContainer" containerID="1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.602746 4809 scope.go:117] "RemoveContainer" containerID="1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035" Dec 05 12:06:22 crc kubenswrapper[4809]: E1205 12:06:22.603209 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035\": container with ID starting with 1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035 not found: ID does not exist" containerID="1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.603241 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035"} err="failed to get container status \"1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035\": rpc error: code = NotFound desc = could not find container \"1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035\": container with ID starting with 1acba0d25f164808c9ec7cbcaca894cac6933f8b3befc3ac2b9c608b9026f035 not found: ID does not exist" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.603267 4809 scope.go:117] "RemoveContainer" containerID="42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3" Dec 05 12:06:22 crc kubenswrapper[4809]: E1205 12:06:22.603549 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3\": container with ID starting with 42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3 not found: ID does not exist" containerID="42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.603589 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3"} err="failed to get container status \"42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3\": rpc error: code = NotFound desc = could not find container \"42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3\": container with ID starting with 42067c146edc81bd8c8d875adb0e4fc8e4487ab0e7db738ca059ad19b6fb0bf3 not found: ID does not exist" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.603610 4809 scope.go:117] "RemoveContainer" containerID="1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85" Dec 05 12:06:22 crc kubenswrapper[4809]: E1205 12:06:22.603826 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85\": container with ID starting with 1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85 not found: ID does not exist" containerID="1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.603852 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85"} err="failed to get container status \"1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85\": rpc error: code = NotFound desc = could not find container \"1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85\": container with ID starting with 1f81d73b504547b456c125cdf9b3ffbdaf7d473a05d546c5cd8476e5439f7a85 not found: ID does not exist" Dec 05 12:06:22 crc kubenswrapper[4809]: I1205 12:06:22.883758 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" path="/var/lib/kubelet/pods/29fd3dc9-7d75-411e-bc60-8843cab48a33/volumes" Dec 05 12:06:23 crc kubenswrapper[4809]: I1205 12:06:23.909060 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:23 crc kubenswrapper[4809]: I1205 12:06:23.909163 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:23 crc kubenswrapper[4809]: I1205 12:06:23.965572 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:24 crc kubenswrapper[4809]: I1205 12:06:24.585010 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:25 crc kubenswrapper[4809]: I1205 12:06:25.570598 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwjfw"] Dec 05 12:06:26 crc kubenswrapper[4809]: I1205 12:06:26.555491 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zwjfw" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="registry-server" containerID="cri-o://18b6f9f42d4546b892ad8b53a7bd41a67543dc3e054572bbbdba2eea8e1a2bfd" gracePeriod=2 Dec 05 12:06:28 crc kubenswrapper[4809]: I1205 12:06:28.574490 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerID="18b6f9f42d4546b892ad8b53a7bd41a67543dc3e054572bbbdba2eea8e1a2bfd" exitCode=0 Dec 05 12:06:28 crc kubenswrapper[4809]: I1205 12:06:28.574600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjfw" event={"ID":"ae8536c1-4129-415f-aabc-7b8bff16962d","Type":"ContainerDied","Data":"18b6f9f42d4546b892ad8b53a7bd41a67543dc3e054572bbbdba2eea8e1a2bfd"} Dec 05 12:06:28 crc kubenswrapper[4809]: I1205 12:06:28.975162 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.064018 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-utilities\") pod \"ae8536c1-4129-415f-aabc-7b8bff16962d\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.064267 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d699\" (UniqueName: \"kubernetes.io/projected/ae8536c1-4129-415f-aabc-7b8bff16962d-kube-api-access-5d699\") pod \"ae8536c1-4129-415f-aabc-7b8bff16962d\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.064306 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-catalog-content\") pod \"ae8536c1-4129-415f-aabc-7b8bff16962d\" (UID: \"ae8536c1-4129-415f-aabc-7b8bff16962d\") " Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.065720 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-utilities" (OuterVolumeSpecName: "utilities") pod "ae8536c1-4129-415f-aabc-7b8bff16962d" (UID: "ae8536c1-4129-415f-aabc-7b8bff16962d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.069340 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8536c1-4129-415f-aabc-7b8bff16962d-kube-api-access-5d699" (OuterVolumeSpecName: "kube-api-access-5d699") pod "ae8536c1-4129-415f-aabc-7b8bff16962d" (UID: "ae8536c1-4129-415f-aabc-7b8bff16962d"). InnerVolumeSpecName "kube-api-access-5d699". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.120796 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae8536c1-4129-415f-aabc-7b8bff16962d" (UID: "ae8536c1-4129-415f-aabc-7b8bff16962d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.165427 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d699\" (UniqueName: \"kubernetes.io/projected/ae8536c1-4129-415f-aabc-7b8bff16962d-kube-api-access-5d699\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.165464 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.165476 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8536c1-4129-415f-aabc-7b8bff16962d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.585433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwjfw" event={"ID":"ae8536c1-4129-415f-aabc-7b8bff16962d","Type":"ContainerDied","Data":"0e04a8c844ab1929c9b72458602a87640d9f2a7c18c90433710e6f73ffe55d8f"} Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.585821 4809 scope.go:117] "RemoveContainer" containerID="18b6f9f42d4546b892ad8b53a7bd41a67543dc3e054572bbbdba2eea8e1a2bfd" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.585970 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwjfw" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.610015 4809 scope.go:117] "RemoveContainer" containerID="6aca686601170362635755f9585e77ff277e2322b6e15c76ec9751c5fdee2550" Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.625822 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwjfw"] Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.633802 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zwjfw"] Dec 05 12:06:29 crc kubenswrapper[4809]: I1205 12:06:29.635908 4809 scope.go:117] "RemoveContainer" containerID="62f824397585fbb9c780b814d38c8125c909e91fd09aea9b1ef3e350af8c1779" Dec 05 12:06:30 crc kubenswrapper[4809]: I1205 12:06:30.881330 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" path="/var/lib/kubelet/pods/ae8536c1-4129-415f-aabc-7b8bff16962d/volumes" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.471572 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-72wwn"] Dec 05 12:06:31 crc kubenswrapper[4809]: E1205 12:06:31.472271 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="registry-server" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472289 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="registry-server" Dec 05 12:06:31 crc kubenswrapper[4809]: E1205 12:06:31.472302 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="extract-content" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472311 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="extract-content" Dec 05 12:06:31 crc kubenswrapper[4809]: E1205 12:06:31.472327 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="extract-utilities" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472336 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="extract-utilities" Dec 05 12:06:31 crc kubenswrapper[4809]: E1205 12:06:31.472363 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="extract-content" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472370 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="extract-content" Dec 05 12:06:31 crc kubenswrapper[4809]: E1205 12:06:31.472380 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="registry-server" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472390 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="registry-server" Dec 05 12:06:31 crc kubenswrapper[4809]: E1205 12:06:31.472408 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="extract-utilities" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472431 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="extract-utilities" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472619 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="29fd3dc9-7d75-411e-bc60-8843cab48a33" containerName="registry-server" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.472664 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8536c1-4129-415f-aabc-7b8bff16962d" containerName="registry-server" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.473656 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.481184 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-72wwn"] Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.601474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-utilities\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.601593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-catalog-content\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.601701 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58pnm\" (UniqueName: \"kubernetes.io/projected/5471084a-6cef-4c99-8931-b5b5acd13bae-kube-api-access-58pnm\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.703473 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-catalog-content\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.703583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58pnm\" (UniqueName: \"kubernetes.io/projected/5471084a-6cef-4c99-8931-b5b5acd13bae-kube-api-access-58pnm\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.703671 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-utilities\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.704032 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-catalog-content\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.704116 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-utilities\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.750757 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58pnm\" (UniqueName: \"kubernetes.io/projected/5471084a-6cef-4c99-8931-b5b5acd13bae-kube-api-access-58pnm\") pod \"certified-operators-72wwn\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:31 crc kubenswrapper[4809]: I1205 12:06:31.790361 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:32 crc kubenswrapper[4809]: I1205 12:06:32.310783 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-72wwn"] Dec 05 12:06:32 crc kubenswrapper[4809]: I1205 12:06:32.617360 4809 generic.go:334] "Generic (PLEG): container finished" podID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerID="00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602" exitCode=0 Dec 05 12:06:32 crc kubenswrapper[4809]: I1205 12:06:32.617421 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72wwn" event={"ID":"5471084a-6cef-4c99-8931-b5b5acd13bae","Type":"ContainerDied","Data":"00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602"} Dec 05 12:06:32 crc kubenswrapper[4809]: I1205 12:06:32.617482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72wwn" event={"ID":"5471084a-6cef-4c99-8931-b5b5acd13bae","Type":"ContainerStarted","Data":"c0c21fbaac36ffabae4df0ac194975c190899a4869a89c26365c183a14599e17"} Dec 05 12:06:33 crc kubenswrapper[4809]: I1205 12:06:33.625688 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72wwn" event={"ID":"5471084a-6cef-4c99-8931-b5b5acd13bae","Type":"ContainerStarted","Data":"6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272"} Dec 05 12:06:34 crc kubenswrapper[4809]: I1205 12:06:34.635032 4809 generic.go:334] "Generic (PLEG): container finished" podID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerID="6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272" exitCode=0 Dec 05 12:06:34 crc kubenswrapper[4809]: I1205 12:06:34.635083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72wwn" event={"ID":"5471084a-6cef-4c99-8931-b5b5acd13bae","Type":"ContainerDied","Data":"6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272"} Dec 05 12:06:35 crc kubenswrapper[4809]: I1205 12:06:35.644461 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72wwn" event={"ID":"5471084a-6cef-4c99-8931-b5b5acd13bae","Type":"ContainerStarted","Data":"ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef"} Dec 05 12:06:35 crc kubenswrapper[4809]: I1205 12:06:35.660650 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-72wwn" podStartSLOduration=2.24944063 podStartE2EDuration="4.660619307s" podCreationTimestamp="2025-12-05 12:06:31 +0000 UTC" firstStartedPulling="2025-12-05 12:06:32.619029299 +0000 UTC m=+3488.010005857" lastFinishedPulling="2025-12-05 12:06:35.030207976 +0000 UTC m=+3490.421184534" observedRunningTime="2025-12-05 12:06:35.659537567 +0000 UTC m=+3491.050514125" watchObservedRunningTime="2025-12-05 12:06:35.660619307 +0000 UTC m=+3491.051595865" Dec 05 12:06:41 crc kubenswrapper[4809]: I1205 12:06:41.790800 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:41 crc kubenswrapper[4809]: I1205 12:06:41.791403 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:41 crc kubenswrapper[4809]: I1205 12:06:41.836062 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:42 crc kubenswrapper[4809]: I1205 12:06:42.741704 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:42 crc kubenswrapper[4809]: I1205 12:06:42.794873 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-72wwn"] Dec 05 12:06:44 crc kubenswrapper[4809]: I1205 12:06:44.705586 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-72wwn" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="registry-server" containerID="cri-o://ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef" gracePeriod=2 Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.576646 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.710025 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58pnm\" (UniqueName: \"kubernetes.io/projected/5471084a-6cef-4c99-8931-b5b5acd13bae-kube-api-access-58pnm\") pod \"5471084a-6cef-4c99-8931-b5b5acd13bae\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.710072 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-catalog-content\") pod \"5471084a-6cef-4c99-8931-b5b5acd13bae\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.710189 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-utilities\") pod \"5471084a-6cef-4c99-8931-b5b5acd13bae\" (UID: \"5471084a-6cef-4c99-8931-b5b5acd13bae\") " Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.711196 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-utilities" (OuterVolumeSpecName: "utilities") pod "5471084a-6cef-4c99-8931-b5b5acd13bae" (UID: "5471084a-6cef-4c99-8931-b5b5acd13bae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.714060 4809 generic.go:334] "Generic (PLEG): container finished" podID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerID="ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef" exitCode=0 Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.714118 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72wwn" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.714127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72wwn" event={"ID":"5471084a-6cef-4c99-8931-b5b5acd13bae","Type":"ContainerDied","Data":"ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef"} Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.714251 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72wwn" event={"ID":"5471084a-6cef-4c99-8931-b5b5acd13bae","Type":"ContainerDied","Data":"c0c21fbaac36ffabae4df0ac194975c190899a4869a89c26365c183a14599e17"} Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.714268 4809 scope.go:117] "RemoveContainer" containerID="ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.716514 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5471084a-6cef-4c99-8931-b5b5acd13bae-kube-api-access-58pnm" (OuterVolumeSpecName: "kube-api-access-58pnm") pod "5471084a-6cef-4c99-8931-b5b5acd13bae" (UID: "5471084a-6cef-4c99-8931-b5b5acd13bae"). InnerVolumeSpecName "kube-api-access-58pnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.751886 4809 scope.go:117] "RemoveContainer" containerID="6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.761396 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5471084a-6cef-4c99-8931-b5b5acd13bae" (UID: "5471084a-6cef-4c99-8931-b5b5acd13bae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.768946 4809 scope.go:117] "RemoveContainer" containerID="00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.789933 4809 scope.go:117] "RemoveContainer" containerID="ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef" Dec 05 12:06:45 crc kubenswrapper[4809]: E1205 12:06:45.790344 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef\": container with ID starting with ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef not found: ID does not exist" containerID="ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.790390 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef"} err="failed to get container status \"ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef\": rpc error: code = NotFound desc = could not find container \"ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef\": container with ID starting with ab0975b63b9ec118dd37fbc9cdf54d3ce2592cc2e9a66953bf531f021e975cef not found: ID does not exist" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.790412 4809 scope.go:117] "RemoveContainer" containerID="6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272" Dec 05 12:06:45 crc kubenswrapper[4809]: E1205 12:06:45.790670 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272\": container with ID starting with 6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272 not found: ID does not exist" containerID="6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.790693 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272"} err="failed to get container status \"6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272\": rpc error: code = NotFound desc = could not find container \"6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272\": container with ID starting with 6bac1938a032243c8ea89aa4eb63f7239ad1fcdf9994f36dbcc942ba9f0b5272 not found: ID does not exist" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.790708 4809 scope.go:117] "RemoveContainer" containerID="00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602" Dec 05 12:06:45 crc kubenswrapper[4809]: E1205 12:06:45.790958 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602\": container with ID starting with 00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602 not found: ID does not exist" containerID="00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.791002 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602"} err="failed to get container status \"00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602\": rpc error: code = NotFound desc = could not find container \"00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602\": container with ID starting with 00caa94ee4139b4c6b62497eb672e0c9b285c4ec54a1a500ddd5b69fe7be5602 not found: ID does not exist" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.811690 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58pnm\" (UniqueName: \"kubernetes.io/projected/5471084a-6cef-4c99-8931-b5b5acd13bae-kube-api-access-58pnm\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.811723 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:45 crc kubenswrapper[4809]: I1205 12:06:45.811747 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5471084a-6cef-4c99-8931-b5b5acd13bae-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:06:46 crc kubenswrapper[4809]: I1205 12:06:46.056366 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-72wwn"] Dec 05 12:06:46 crc kubenswrapper[4809]: I1205 12:06:46.068901 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-72wwn"] Dec 05 12:06:46 crc kubenswrapper[4809]: I1205 12:06:46.884957 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" path="/var/lib/kubelet/pods/5471084a-6cef-4c99-8931-b5b5acd13bae/volumes" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.759528 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t7nw9"] Dec 05 12:07:16 crc kubenswrapper[4809]: E1205 12:07:16.760464 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="registry-server" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.760485 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="registry-server" Dec 05 12:07:16 crc kubenswrapper[4809]: E1205 12:07:16.760503 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="extract-content" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.760514 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="extract-content" Dec 05 12:07:16 crc kubenswrapper[4809]: E1205 12:07:16.760542 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="extract-utilities" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.760552 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="extract-utilities" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.760762 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5471084a-6cef-4c99-8931-b5b5acd13bae" containerName="registry-server" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.762150 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.772471 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7nw9"] Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.872125 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-catalog-content\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.872185 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z7zb\" (UniqueName: \"kubernetes.io/projected/266e45d0-cbae-4e95-b5e2-92095587201b-kube-api-access-9z7zb\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.872249 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-utilities\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.973876 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-catalog-content\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.973953 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z7zb\" (UniqueName: \"kubernetes.io/projected/266e45d0-cbae-4e95-b5e2-92095587201b-kube-api-access-9z7zb\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.973993 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-utilities\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.974555 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-catalog-content\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.974609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-utilities\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:16 crc kubenswrapper[4809]: I1205 12:07:16.993827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z7zb\" (UniqueName: \"kubernetes.io/projected/266e45d0-cbae-4e95-b5e2-92095587201b-kube-api-access-9z7zb\") pod \"redhat-operators-t7nw9\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:17 crc kubenswrapper[4809]: I1205 12:07:17.129581 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:17 crc kubenswrapper[4809]: I1205 12:07:17.555088 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7nw9"] Dec 05 12:07:17 crc kubenswrapper[4809]: I1205 12:07:17.945019 4809 generic.go:334] "Generic (PLEG): container finished" podID="266e45d0-cbae-4e95-b5e2-92095587201b" containerID="860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599" exitCode=0 Dec 05 12:07:17 crc kubenswrapper[4809]: I1205 12:07:17.945112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7nw9" event={"ID":"266e45d0-cbae-4e95-b5e2-92095587201b","Type":"ContainerDied","Data":"860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599"} Dec 05 12:07:17 crc kubenswrapper[4809]: I1205 12:07:17.945309 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7nw9" event={"ID":"266e45d0-cbae-4e95-b5e2-92095587201b","Type":"ContainerStarted","Data":"fe81ec13bad9f532c6d6c90a356bc78b7db75a88dacdbc7d00c53e0015757775"} Dec 05 12:07:18 crc kubenswrapper[4809]: I1205 12:07:18.955555 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7nw9" event={"ID":"266e45d0-cbae-4e95-b5e2-92095587201b","Type":"ContainerStarted","Data":"fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef"} Dec 05 12:07:19 crc kubenswrapper[4809]: I1205 12:07:19.964429 4809 generic.go:334] "Generic (PLEG): container finished" podID="266e45d0-cbae-4e95-b5e2-92095587201b" containerID="fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef" exitCode=0 Dec 05 12:07:19 crc kubenswrapper[4809]: I1205 12:07:19.964480 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7nw9" event={"ID":"266e45d0-cbae-4e95-b5e2-92095587201b","Type":"ContainerDied","Data":"fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef"} Dec 05 12:07:20 crc kubenswrapper[4809]: I1205 12:07:20.973409 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7nw9" event={"ID":"266e45d0-cbae-4e95-b5e2-92095587201b","Type":"ContainerStarted","Data":"a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee"} Dec 05 12:07:20 crc kubenswrapper[4809]: I1205 12:07:20.991509 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t7nw9" podStartSLOduration=2.478182387 podStartE2EDuration="4.99149136s" podCreationTimestamp="2025-12-05 12:07:16 +0000 UTC" firstStartedPulling="2025-12-05 12:07:17.946840352 +0000 UTC m=+3533.337816910" lastFinishedPulling="2025-12-05 12:07:20.460149335 +0000 UTC m=+3535.851125883" observedRunningTime="2025-12-05 12:07:20.989081485 +0000 UTC m=+3536.380058053" watchObservedRunningTime="2025-12-05 12:07:20.99149136 +0000 UTC m=+3536.382467918" Dec 05 12:07:27 crc kubenswrapper[4809]: I1205 12:07:27.130356 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:27 crc kubenswrapper[4809]: I1205 12:07:27.131064 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:27 crc kubenswrapper[4809]: I1205 12:07:27.186289 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:28 crc kubenswrapper[4809]: I1205 12:07:28.071862 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:28 crc kubenswrapper[4809]: I1205 12:07:28.116122 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7nw9"] Dec 05 12:07:30 crc kubenswrapper[4809]: I1205 12:07:30.046687 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t7nw9" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="registry-server" containerID="cri-o://a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee" gracePeriod=2 Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.013951 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.074522 4809 generic.go:334] "Generic (PLEG): container finished" podID="266e45d0-cbae-4e95-b5e2-92095587201b" containerID="a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee" exitCode=0 Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.074561 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7nw9" event={"ID":"266e45d0-cbae-4e95-b5e2-92095587201b","Type":"ContainerDied","Data":"a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee"} Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.074592 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7nw9" event={"ID":"266e45d0-cbae-4e95-b5e2-92095587201b","Type":"ContainerDied","Data":"fe81ec13bad9f532c6d6c90a356bc78b7db75a88dacdbc7d00c53e0015757775"} Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.074614 4809 scope.go:117] "RemoveContainer" containerID="a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.074613 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7nw9" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.099246 4809 scope.go:117] "RemoveContainer" containerID="fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.123435 4809 scope.go:117] "RemoveContainer" containerID="860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.152949 4809 scope.go:117] "RemoveContainer" containerID="a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee" Dec 05 12:07:33 crc kubenswrapper[4809]: E1205 12:07:33.153605 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee\": container with ID starting with a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee not found: ID does not exist" containerID="a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.153909 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee"} err="failed to get container status \"a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee\": rpc error: code = NotFound desc = could not find container \"a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee\": container with ID starting with a3d67b2f49ef2c339af2959532ed8fb3f75f861319942ca8ac1ab684abaa00ee not found: ID does not exist" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.153949 4809 scope.go:117] "RemoveContainer" containerID="fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef" Dec 05 12:07:33 crc kubenswrapper[4809]: E1205 12:07:33.154490 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef\": container with ID starting with fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef not found: ID does not exist" containerID="fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.154515 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef"} err="failed to get container status \"fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef\": rpc error: code = NotFound desc = could not find container \"fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef\": container with ID starting with fe763b5bb6881c763a2928caa433aa09fdc9eb19322a6dc2dd3540d53e978aef not found: ID does not exist" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.154534 4809 scope.go:117] "RemoveContainer" containerID="860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599" Dec 05 12:07:33 crc kubenswrapper[4809]: E1205 12:07:33.154901 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599\": container with ID starting with 860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599 not found: ID does not exist" containerID="860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.154933 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599"} err="failed to get container status \"860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599\": rpc error: code = NotFound desc = could not find container \"860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599\": container with ID starting with 860dbecc8aba5e050ed9f88d8d07985d0b9b69ba3b30036fca86b6ddacd8c599 not found: ID does not exist" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.206063 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-utilities\") pod \"266e45d0-cbae-4e95-b5e2-92095587201b\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.206222 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z7zb\" (UniqueName: \"kubernetes.io/projected/266e45d0-cbae-4e95-b5e2-92095587201b-kube-api-access-9z7zb\") pod \"266e45d0-cbae-4e95-b5e2-92095587201b\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.206260 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-catalog-content\") pod \"266e45d0-cbae-4e95-b5e2-92095587201b\" (UID: \"266e45d0-cbae-4e95-b5e2-92095587201b\") " Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.206972 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-utilities" (OuterVolumeSpecName: "utilities") pod "266e45d0-cbae-4e95-b5e2-92095587201b" (UID: "266e45d0-cbae-4e95-b5e2-92095587201b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.208157 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.213020 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/266e45d0-cbae-4e95-b5e2-92095587201b-kube-api-access-9z7zb" (OuterVolumeSpecName: "kube-api-access-9z7zb") pod "266e45d0-cbae-4e95-b5e2-92095587201b" (UID: "266e45d0-cbae-4e95-b5e2-92095587201b"). InnerVolumeSpecName "kube-api-access-9z7zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.309467 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z7zb\" (UniqueName: \"kubernetes.io/projected/266e45d0-cbae-4e95-b5e2-92095587201b-kube-api-access-9z7zb\") on node \"crc\" DevicePath \"\"" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.340202 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "266e45d0-cbae-4e95-b5e2-92095587201b" (UID: "266e45d0-cbae-4e95-b5e2-92095587201b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.412123 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/266e45d0-cbae-4e95-b5e2-92095587201b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.413978 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7nw9"] Dec 05 12:07:33 crc kubenswrapper[4809]: I1205 12:07:33.419292 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t7nw9"] Dec 05 12:07:34 crc kubenswrapper[4809]: I1205 12:07:34.882669 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" path="/var/lib/kubelet/pods/266e45d0-cbae-4e95-b5e2-92095587201b/volumes" Dec 05 12:07:44 crc kubenswrapper[4809]: I1205 12:07:44.046715 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:07:44 crc kubenswrapper[4809]: I1205 12:07:44.047321 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:08:14 crc kubenswrapper[4809]: I1205 12:08:14.047167 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:08:14 crc kubenswrapper[4809]: I1205 12:08:14.047770 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.047819 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.048416 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.048474 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.049201 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.049271 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" gracePeriod=600 Dec 05 12:08:44 crc kubenswrapper[4809]: E1205 12:08:44.172195 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.633061 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" exitCode=0 Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.633103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2"} Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.633133 4809 scope.go:117] "RemoveContainer" containerID="d61be0cb10852d25dee858d34fd89106d48f1e911a9ce1010259ac7fc3a322f9" Dec 05 12:08:44 crc kubenswrapper[4809]: I1205 12:08:44.633662 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:08:44 crc kubenswrapper[4809]: E1205 12:08:44.633916 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:08:57 crc kubenswrapper[4809]: I1205 12:08:57.872670 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:08:57 crc kubenswrapper[4809]: E1205 12:08:57.874501 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:09:08 crc kubenswrapper[4809]: I1205 12:09:08.872948 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:09:08 crc kubenswrapper[4809]: E1205 12:09:08.873675 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:09:19 crc kubenswrapper[4809]: I1205 12:09:19.872402 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:09:19 crc kubenswrapper[4809]: E1205 12:09:19.872950 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:09:32 crc kubenswrapper[4809]: I1205 12:09:32.872554 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:09:32 crc kubenswrapper[4809]: E1205 12:09:32.873383 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:09:46 crc kubenswrapper[4809]: I1205 12:09:46.872813 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:09:46 crc kubenswrapper[4809]: E1205 12:09:46.873756 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:10:00 crc kubenswrapper[4809]: I1205 12:10:00.872544 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:10:00 crc kubenswrapper[4809]: E1205 12:10:00.873766 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:10:13 crc kubenswrapper[4809]: I1205 12:10:13.872188 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:10:13 crc kubenswrapper[4809]: E1205 12:10:13.873099 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:10:25 crc kubenswrapper[4809]: I1205 12:10:25.871983 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:10:25 crc kubenswrapper[4809]: E1205 12:10:25.872661 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:10:37 crc kubenswrapper[4809]: I1205 12:10:37.872505 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:10:37 crc kubenswrapper[4809]: E1205 12:10:37.873370 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:10:48 crc kubenswrapper[4809]: I1205 12:10:48.873137 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:10:48 crc kubenswrapper[4809]: E1205 12:10:48.873761 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:11:03 crc kubenswrapper[4809]: I1205 12:11:03.873129 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:11:03 crc kubenswrapper[4809]: E1205 12:11:03.874011 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:11:14 crc kubenswrapper[4809]: I1205 12:11:14.875779 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:11:14 crc kubenswrapper[4809]: E1205 12:11:14.877608 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:11:29 crc kubenswrapper[4809]: I1205 12:11:29.872419 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:11:29 crc kubenswrapper[4809]: E1205 12:11:29.873375 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:11:41 crc kubenswrapper[4809]: I1205 12:11:41.873737 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:11:41 crc kubenswrapper[4809]: E1205 12:11:41.874675 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:11:55 crc kubenswrapper[4809]: I1205 12:11:55.872735 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:11:55 crc kubenswrapper[4809]: E1205 12:11:55.873930 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:12:10 crc kubenswrapper[4809]: I1205 12:12:10.873374 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:12:10 crc kubenswrapper[4809]: E1205 12:12:10.874237 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:12:23 crc kubenswrapper[4809]: I1205 12:12:23.872252 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:12:23 crc kubenswrapper[4809]: E1205 12:12:23.873030 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:12:38 crc kubenswrapper[4809]: I1205 12:12:38.874133 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:12:38 crc kubenswrapper[4809]: E1205 12:12:38.875090 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:12:53 crc kubenswrapper[4809]: I1205 12:12:53.872488 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:12:53 crc kubenswrapper[4809]: E1205 12:12:53.873453 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:13:04 crc kubenswrapper[4809]: I1205 12:13:04.879131 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:13:04 crc kubenswrapper[4809]: E1205 12:13:04.879707 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:13:18 crc kubenswrapper[4809]: I1205 12:13:18.872174 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:13:18 crc kubenswrapper[4809]: E1205 12:13:18.873445 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:13:32 crc kubenswrapper[4809]: I1205 12:13:32.872804 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:13:32 crc kubenswrapper[4809]: E1205 12:13:32.873760 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:13:43 crc kubenswrapper[4809]: I1205 12:13:43.872230 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:13:43 crc kubenswrapper[4809]: E1205 12:13:43.873114 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:13:54 crc kubenswrapper[4809]: I1205 12:13:54.876645 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:13:56 crc kubenswrapper[4809]: I1205 12:13:56.038652 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"d9a9dd09031827650d11d1b983fac8b8ba6eeb9c83c05f84175ab691223ff77f"} Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.177053 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt"] Dec 05 12:15:00 crc kubenswrapper[4809]: E1205 12:15:00.178017 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="registry-server" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.178034 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="registry-server" Dec 05 12:15:00 crc kubenswrapper[4809]: E1205 12:15:00.178048 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="extract-utilities" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.178055 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="extract-utilities" Dec 05 12:15:00 crc kubenswrapper[4809]: E1205 12:15:00.178071 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="extract-content" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.178077 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="extract-content" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.178217 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="266e45d0-cbae-4e95-b5e2-92095587201b" containerName="registry-server" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.178758 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.180992 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.181329 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.224995 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt"] Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.373575 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d40cc40-03ff-48fd-9a15-04b68ce4b435-secret-volume\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.373711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45rc9\" (UniqueName: \"kubernetes.io/projected/1d40cc40-03ff-48fd-9a15-04b68ce4b435-kube-api-access-45rc9\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.373736 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d40cc40-03ff-48fd-9a15-04b68ce4b435-config-volume\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.474578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d40cc40-03ff-48fd-9a15-04b68ce4b435-secret-volume\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.474694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45rc9\" (UniqueName: \"kubernetes.io/projected/1d40cc40-03ff-48fd-9a15-04b68ce4b435-kube-api-access-45rc9\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.474719 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d40cc40-03ff-48fd-9a15-04b68ce4b435-config-volume\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.475797 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d40cc40-03ff-48fd-9a15-04b68ce4b435-config-volume\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.480968 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d40cc40-03ff-48fd-9a15-04b68ce4b435-secret-volume\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.492067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45rc9\" (UniqueName: \"kubernetes.io/projected/1d40cc40-03ff-48fd-9a15-04b68ce4b435-kube-api-access-45rc9\") pod \"collect-profiles-29415615-6tpwt\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.499784 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:00 crc kubenswrapper[4809]: I1205 12:15:00.946041 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt"] Dec 05 12:15:01 crc kubenswrapper[4809]: I1205 12:15:01.496062 4809 generic.go:334] "Generic (PLEG): container finished" podID="1d40cc40-03ff-48fd-9a15-04b68ce4b435" containerID="589c2763c22f289f78f62648b0c5a53b6a47489b8720fd173824cdb148050e16" exitCode=0 Dec 05 12:15:01 crc kubenswrapper[4809]: I1205 12:15:01.496180 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" event={"ID":"1d40cc40-03ff-48fd-9a15-04b68ce4b435","Type":"ContainerDied","Data":"589c2763c22f289f78f62648b0c5a53b6a47489b8720fd173824cdb148050e16"} Dec 05 12:15:01 crc kubenswrapper[4809]: I1205 12:15:01.496381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" event={"ID":"1d40cc40-03ff-48fd-9a15-04b68ce4b435","Type":"ContainerStarted","Data":"8769ff8b0dea5363dc113755b813b979a8456e3a5dd208dfa5f7f59c41153509"} Dec 05 12:15:02 crc kubenswrapper[4809]: I1205 12:15:02.872191 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.026618 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d40cc40-03ff-48fd-9a15-04b68ce4b435-secret-volume\") pod \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.026745 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d40cc40-03ff-48fd-9a15-04b68ce4b435-config-volume\") pod \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.026898 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45rc9\" (UniqueName: \"kubernetes.io/projected/1d40cc40-03ff-48fd-9a15-04b68ce4b435-kube-api-access-45rc9\") pod \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\" (UID: \"1d40cc40-03ff-48fd-9a15-04b68ce4b435\") " Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.028870 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d40cc40-03ff-48fd-9a15-04b68ce4b435-config-volume" (OuterVolumeSpecName: "config-volume") pod "1d40cc40-03ff-48fd-9a15-04b68ce4b435" (UID: "1d40cc40-03ff-48fd-9a15-04b68ce4b435"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.032874 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d40cc40-03ff-48fd-9a15-04b68ce4b435-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1d40cc40-03ff-48fd-9a15-04b68ce4b435" (UID: "1d40cc40-03ff-48fd-9a15-04b68ce4b435"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.033862 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d40cc40-03ff-48fd-9a15-04b68ce4b435-kube-api-access-45rc9" (OuterVolumeSpecName: "kube-api-access-45rc9") pod "1d40cc40-03ff-48fd-9a15-04b68ce4b435" (UID: "1d40cc40-03ff-48fd-9a15-04b68ce4b435"). InnerVolumeSpecName "kube-api-access-45rc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.128425 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45rc9\" (UniqueName: \"kubernetes.io/projected/1d40cc40-03ff-48fd-9a15-04b68ce4b435-kube-api-access-45rc9\") on node \"crc\" DevicePath \"\"" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.128468 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d40cc40-03ff-48fd-9a15-04b68ce4b435-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.128491 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d40cc40-03ff-48fd-9a15-04b68ce4b435-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.510489 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" event={"ID":"1d40cc40-03ff-48fd-9a15-04b68ce4b435","Type":"ContainerDied","Data":"8769ff8b0dea5363dc113755b813b979a8456e3a5dd208dfa5f7f59c41153509"} Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.510555 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8769ff8b0dea5363dc113755b813b979a8456e3a5dd208dfa5f7f59c41153509" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.510624 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt" Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.937396 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr"] Dec 05 12:15:03 crc kubenswrapper[4809]: I1205 12:15:03.944003 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-drknr"] Dec 05 12:15:04 crc kubenswrapper[4809]: I1205 12:15:04.881021 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ee7bf0-2a51-4926-b47d-c06c792c1fe9" path="/var/lib/kubelet/pods/19ee7bf0-2a51-4926-b47d-c06c792c1fe9/volumes" Dec 05 12:15:33 crc kubenswrapper[4809]: I1205 12:15:33.678069 4809 scope.go:117] "RemoveContainer" containerID="37f4400738a589859e4d3607ebf099a53bd2eb5843646f4ec28456a932d2123d" Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.945471 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-clfcv"] Dec 05 12:16:07 crc kubenswrapper[4809]: E1205 12:16:07.946553 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d40cc40-03ff-48fd-9a15-04b68ce4b435" containerName="collect-profiles" Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.946567 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d40cc40-03ff-48fd-9a15-04b68ce4b435" containerName="collect-profiles" Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.946765 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d40cc40-03ff-48fd-9a15-04b68ce4b435" containerName="collect-profiles" Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.947726 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.964699 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-clfcv"] Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.971622 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-catalog-content\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.971713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-utilities\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:07 crc kubenswrapper[4809]: I1205 12:16:07.971768 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twf6d\" (UniqueName: \"kubernetes.io/projected/1a13fed6-cd34-4288-b5de-a7cab6d507c7-kube-api-access-twf6d\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.072691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-catalog-content\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.072751 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-utilities\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.072805 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twf6d\" (UniqueName: \"kubernetes.io/projected/1a13fed6-cd34-4288-b5de-a7cab6d507c7-kube-api-access-twf6d\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.073337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-catalog-content\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.073376 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-utilities\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.092055 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twf6d\" (UniqueName: \"kubernetes.io/projected/1a13fed6-cd34-4288-b5de-a7cab6d507c7-kube-api-access-twf6d\") pod \"redhat-marketplace-clfcv\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.268172 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:08 crc kubenswrapper[4809]: I1205 12:16:08.771718 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-clfcv"] Dec 05 12:16:09 crc kubenswrapper[4809]: W1205 12:16:09.181736 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a13fed6_cd34_4288_b5de_a7cab6d507c7.slice/crio-84d8e79fb08d0be9189dc255ae7ebd1154280e30114473df620407a44c299c5d WatchSource:0}: Error finding container 84d8e79fb08d0be9189dc255ae7ebd1154280e30114473df620407a44c299c5d: Status 404 returned error can't find the container with id 84d8e79fb08d0be9189dc255ae7ebd1154280e30114473df620407a44c299c5d Dec 05 12:16:09 crc kubenswrapper[4809]: I1205 12:16:09.999405 4809 generic.go:334] "Generic (PLEG): container finished" podID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerID="7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e" exitCode=0 Dec 05 12:16:09 crc kubenswrapper[4809]: I1205 12:16:09.999703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clfcv" event={"ID":"1a13fed6-cd34-4288-b5de-a7cab6d507c7","Type":"ContainerDied","Data":"7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e"} Dec 05 12:16:10 crc kubenswrapper[4809]: I1205 12:16:09.999733 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clfcv" event={"ID":"1a13fed6-cd34-4288-b5de-a7cab6d507c7","Type":"ContainerStarted","Data":"84d8e79fb08d0be9189dc255ae7ebd1154280e30114473df620407a44c299c5d"} Dec 05 12:16:10 crc kubenswrapper[4809]: I1205 12:16:10.002205 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:16:11 crc kubenswrapper[4809]: I1205 12:16:11.010984 4809 generic.go:334] "Generic (PLEG): container finished" podID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerID="b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0" exitCode=0 Dec 05 12:16:11 crc kubenswrapper[4809]: I1205 12:16:11.011056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clfcv" event={"ID":"1a13fed6-cd34-4288-b5de-a7cab6d507c7","Type":"ContainerDied","Data":"b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0"} Dec 05 12:16:12 crc kubenswrapper[4809]: I1205 12:16:12.020040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clfcv" event={"ID":"1a13fed6-cd34-4288-b5de-a7cab6d507c7","Type":"ContainerStarted","Data":"00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe"} Dec 05 12:16:12 crc kubenswrapper[4809]: I1205 12:16:12.040523 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-clfcv" podStartSLOduration=3.593091116 podStartE2EDuration="5.040504747s" podCreationTimestamp="2025-12-05 12:16:07 +0000 UTC" firstStartedPulling="2025-12-05 12:16:10.001831801 +0000 UTC m=+4065.392808359" lastFinishedPulling="2025-12-05 12:16:11.449245412 +0000 UTC m=+4066.840221990" observedRunningTime="2025-12-05 12:16:12.037215109 +0000 UTC m=+4067.428191687" watchObservedRunningTime="2025-12-05 12:16:12.040504747 +0000 UTC m=+4067.431481305" Dec 05 12:16:14 crc kubenswrapper[4809]: I1205 12:16:14.047447 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:16:14 crc kubenswrapper[4809]: I1205 12:16:14.047536 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:16:18 crc kubenswrapper[4809]: I1205 12:16:18.269522 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:18 crc kubenswrapper[4809]: I1205 12:16:18.270076 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:18 crc kubenswrapper[4809]: I1205 12:16:18.316225 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:19 crc kubenswrapper[4809]: I1205 12:16:19.117968 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:21 crc kubenswrapper[4809]: I1205 12:16:21.936964 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-clfcv"] Dec 05 12:16:21 crc kubenswrapper[4809]: I1205 12:16:21.937535 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-clfcv" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="registry-server" containerID="cri-o://00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe" gracePeriod=2 Dec 05 12:16:22 crc kubenswrapper[4809]: I1205 12:16:22.815743 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:22 crc kubenswrapper[4809]: I1205 12:16:22.994201 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-utilities\") pod \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " Dec 05 12:16:22 crc kubenswrapper[4809]: I1205 12:16:22.994315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-catalog-content\") pod \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " Dec 05 12:16:22 crc kubenswrapper[4809]: I1205 12:16:22.994461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twf6d\" (UniqueName: \"kubernetes.io/projected/1a13fed6-cd34-4288-b5de-a7cab6d507c7-kube-api-access-twf6d\") pod \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\" (UID: \"1a13fed6-cd34-4288-b5de-a7cab6d507c7\") " Dec 05 12:16:22 crc kubenswrapper[4809]: I1205 12:16:22.995317 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-utilities" (OuterVolumeSpecName: "utilities") pod "1a13fed6-cd34-4288-b5de-a7cab6d507c7" (UID: "1a13fed6-cd34-4288-b5de-a7cab6d507c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.001948 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a13fed6-cd34-4288-b5de-a7cab6d507c7-kube-api-access-twf6d" (OuterVolumeSpecName: "kube-api-access-twf6d") pod "1a13fed6-cd34-4288-b5de-a7cab6d507c7" (UID: "1a13fed6-cd34-4288-b5de-a7cab6d507c7"). InnerVolumeSpecName "kube-api-access-twf6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.012903 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a13fed6-cd34-4288-b5de-a7cab6d507c7" (UID: "1a13fed6-cd34-4288-b5de-a7cab6d507c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.095869 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.095917 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a13fed6-cd34-4288-b5de-a7cab6d507c7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.095933 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twf6d\" (UniqueName: \"kubernetes.io/projected/1a13fed6-cd34-4288-b5de-a7cab6d507c7-kube-api-access-twf6d\") on node \"crc\" DevicePath \"\"" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.106147 4809 generic.go:334] "Generic (PLEG): container finished" podID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerID="00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe" exitCode=0 Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.106188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clfcv" event={"ID":"1a13fed6-cd34-4288-b5de-a7cab6d507c7","Type":"ContainerDied","Data":"00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe"} Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.106215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clfcv" event={"ID":"1a13fed6-cd34-4288-b5de-a7cab6d507c7","Type":"ContainerDied","Data":"84d8e79fb08d0be9189dc255ae7ebd1154280e30114473df620407a44c299c5d"} Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.106233 4809 scope.go:117] "RemoveContainer" containerID="00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.106238 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clfcv" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.143959 4809 scope.go:117] "RemoveContainer" containerID="b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.149003 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-clfcv"] Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.155252 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-clfcv"] Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.175437 4809 scope.go:117] "RemoveContainer" containerID="7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.194808 4809 scope.go:117] "RemoveContainer" containerID="00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe" Dec 05 12:16:23 crc kubenswrapper[4809]: E1205 12:16:23.195233 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe\": container with ID starting with 00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe not found: ID does not exist" containerID="00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.195275 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe"} err="failed to get container status \"00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe\": rpc error: code = NotFound desc = could not find container \"00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe\": container with ID starting with 00dece6e7bb89c44e6f845f2b29516f46bc8fd8914c6e0c2d9ac8bcb88d711fe not found: ID does not exist" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.195305 4809 scope.go:117] "RemoveContainer" containerID="b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0" Dec 05 12:16:23 crc kubenswrapper[4809]: E1205 12:16:23.195773 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0\": container with ID starting with b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0 not found: ID does not exist" containerID="b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.195821 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0"} err="failed to get container status \"b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0\": rpc error: code = NotFound desc = could not find container \"b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0\": container with ID starting with b5b99e880f05be178f43b4c5bdd00cae65131bbde01e3e705070c94226d1d9c0 not found: ID does not exist" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.195853 4809 scope.go:117] "RemoveContainer" containerID="7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e" Dec 05 12:16:23 crc kubenswrapper[4809]: E1205 12:16:23.196137 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e\": container with ID starting with 7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e not found: ID does not exist" containerID="7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e" Dec 05 12:16:23 crc kubenswrapper[4809]: I1205 12:16:23.196165 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e"} err="failed to get container status \"7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e\": rpc error: code = NotFound desc = could not find container \"7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e\": container with ID starting with 7dcab6ad900cec6fc74fc519bcc36029e8a183320b1a50ab6d11f7c72fda963e not found: ID does not exist" Dec 05 12:16:24 crc kubenswrapper[4809]: I1205 12:16:24.885982 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" path="/var/lib/kubelet/pods/1a13fed6-cd34-4288-b5de-a7cab6d507c7/volumes" Dec 05 12:16:44 crc kubenswrapper[4809]: I1205 12:16:44.047839 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:16:44 crc kubenswrapper[4809]: I1205 12:16:44.049882 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.046786 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.047340 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.047396 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.048047 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9a9dd09031827650d11d1b983fac8b8ba6eeb9c83c05f84175ab691223ff77f"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.048129 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://d9a9dd09031827650d11d1b983fac8b8ba6eeb9c83c05f84175ab691223ff77f" gracePeriod=600 Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.603622 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="d9a9dd09031827650d11d1b983fac8b8ba6eeb9c83c05f84175ab691223ff77f" exitCode=0 Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.603860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"d9a9dd09031827650d11d1b983fac8b8ba6eeb9c83c05f84175ab691223ff77f"} Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.604054 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8"} Dec 05 12:17:14 crc kubenswrapper[4809]: I1205 12:17:14.604085 4809 scope.go:117] "RemoveContainer" containerID="e0af892aaf6c5cc92b6149c15566f2fea11a94db5cea93b217d7746ee5b5f0c2" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.738397 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rdgst"] Dec 05 12:17:19 crc kubenswrapper[4809]: E1205 12:17:19.739074 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="extract-utilities" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.739094 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="extract-utilities" Dec 05 12:17:19 crc kubenswrapper[4809]: E1205 12:17:19.739105 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="extract-content" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.739113 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="extract-content" Dec 05 12:17:19 crc kubenswrapper[4809]: E1205 12:17:19.739129 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="registry-server" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.739137 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="registry-server" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.739312 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a13fed6-cd34-4288-b5de-a7cab6d507c7" containerName="registry-server" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.740502 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.759243 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdgst"] Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.852712 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-utilities\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.852760 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-catalog-content\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.853001 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcnk8\" (UniqueName: \"kubernetes.io/projected/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-kube-api-access-dcnk8\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.954195 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcnk8\" (UniqueName: \"kubernetes.io/projected/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-kube-api-access-dcnk8\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.954344 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-utilities\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.954383 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-catalog-content\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.954893 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-catalog-content\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.954920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-utilities\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:19 crc kubenswrapper[4809]: I1205 12:17:19.979947 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcnk8\" (UniqueName: \"kubernetes.io/projected/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-kube-api-access-dcnk8\") pod \"certified-operators-rdgst\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:20 crc kubenswrapper[4809]: I1205 12:17:20.067561 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:20 crc kubenswrapper[4809]: I1205 12:17:20.537747 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdgst"] Dec 05 12:17:20 crc kubenswrapper[4809]: I1205 12:17:20.648697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdgst" event={"ID":"e4ced3d5-2952-4344-89d8-4bbf707bd5f0","Type":"ContainerStarted","Data":"2be1e8db084491609c35e56c65cefb8b6b1160a911f8e116ae7389368ced0069"} Dec 05 12:17:21 crc kubenswrapper[4809]: I1205 12:17:21.659797 4809 generic.go:334] "Generic (PLEG): container finished" podID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerID="419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f" exitCode=0 Dec 05 12:17:21 crc kubenswrapper[4809]: I1205 12:17:21.660034 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdgst" event={"ID":"e4ced3d5-2952-4344-89d8-4bbf707bd5f0","Type":"ContainerDied","Data":"419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f"} Dec 05 12:17:23 crc kubenswrapper[4809]: I1205 12:17:23.676814 4809 generic.go:334] "Generic (PLEG): container finished" podID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerID="0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba" exitCode=0 Dec 05 12:17:23 crc kubenswrapper[4809]: I1205 12:17:23.676900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdgst" event={"ID":"e4ced3d5-2952-4344-89d8-4bbf707bd5f0","Type":"ContainerDied","Data":"0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba"} Dec 05 12:17:24 crc kubenswrapper[4809]: I1205 12:17:24.688455 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdgst" event={"ID":"e4ced3d5-2952-4344-89d8-4bbf707bd5f0","Type":"ContainerStarted","Data":"b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75"} Dec 05 12:17:24 crc kubenswrapper[4809]: I1205 12:17:24.714698 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rdgst" podStartSLOduration=3.30287737 podStartE2EDuration="5.714677166s" podCreationTimestamp="2025-12-05 12:17:19 +0000 UTC" firstStartedPulling="2025-12-05 12:17:21.662825145 +0000 UTC m=+4137.053801713" lastFinishedPulling="2025-12-05 12:17:24.074624951 +0000 UTC m=+4139.465601509" observedRunningTime="2025-12-05 12:17:24.706595089 +0000 UTC m=+4140.097571657" watchObservedRunningTime="2025-12-05 12:17:24.714677166 +0000 UTC m=+4140.105653724" Dec 05 12:17:30 crc kubenswrapper[4809]: I1205 12:17:30.068221 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:30 crc kubenswrapper[4809]: I1205 12:17:30.068562 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:30 crc kubenswrapper[4809]: I1205 12:17:30.305901 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:30 crc kubenswrapper[4809]: I1205 12:17:30.785921 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:30 crc kubenswrapper[4809]: I1205 12:17:30.836341 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdgst"] Dec 05 12:17:32 crc kubenswrapper[4809]: I1205 12:17:32.745800 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rdgst" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="registry-server" containerID="cri-o://b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75" gracePeriod=2 Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.108144 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.145878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcnk8\" (UniqueName: \"kubernetes.io/projected/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-kube-api-access-dcnk8\") pod \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.145989 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-catalog-content\") pod \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.146153 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-utilities\") pod \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\" (UID: \"e4ced3d5-2952-4344-89d8-4bbf707bd5f0\") " Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.148026 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-utilities" (OuterVolumeSpecName: "utilities") pod "e4ced3d5-2952-4344-89d8-4bbf707bd5f0" (UID: "e4ced3d5-2952-4344-89d8-4bbf707bd5f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.153322 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-kube-api-access-dcnk8" (OuterVolumeSpecName: "kube-api-access-dcnk8") pod "e4ced3d5-2952-4344-89d8-4bbf707bd5f0" (UID: "e4ced3d5-2952-4344-89d8-4bbf707bd5f0"). InnerVolumeSpecName "kube-api-access-dcnk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.248133 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.248178 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcnk8\" (UniqueName: \"kubernetes.io/projected/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-kube-api-access-dcnk8\") on node \"crc\" DevicePath \"\"" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.455122 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4ced3d5-2952-4344-89d8-4bbf707bd5f0" (UID: "e4ced3d5-2952-4344-89d8-4bbf707bd5f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.552774 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4ced3d5-2952-4344-89d8-4bbf707bd5f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.757486 4809 generic.go:334] "Generic (PLEG): container finished" podID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerID="b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75" exitCode=0 Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.757646 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdgst" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.757647 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdgst" event={"ID":"e4ced3d5-2952-4344-89d8-4bbf707bd5f0","Type":"ContainerDied","Data":"b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75"} Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.759410 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdgst" event={"ID":"e4ced3d5-2952-4344-89d8-4bbf707bd5f0","Type":"ContainerDied","Data":"2be1e8db084491609c35e56c65cefb8b6b1160a911f8e116ae7389368ced0069"} Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.759445 4809 scope.go:117] "RemoveContainer" containerID="b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75" Dec 05 12:17:33 crc kubenswrapper[4809]: E1205 12:17:33.763374 4809 kuberuntime_gc.go:389] "Failed to remove container log dead symlink" err="remove /var/log/containers/certified-operators-rdgst_openshift-marketplace_registry-server-b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75.log: no such file or directory" path="/var/log/containers/certified-operators-rdgst_openshift-marketplace_registry-server-b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75.log" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.777975 4809 scope.go:117] "RemoveContainer" containerID="0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.799433 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdgst"] Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.806943 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rdgst"] Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.823823 4809 scope.go:117] "RemoveContainer" containerID="419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.853487 4809 scope.go:117] "RemoveContainer" containerID="b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75" Dec 05 12:17:33 crc kubenswrapper[4809]: E1205 12:17:33.854090 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75\": container with ID starting with b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75 not found: ID does not exist" containerID="b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.854151 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75"} err="failed to get container status \"b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75\": rpc error: code = NotFound desc = could not find container \"b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75\": container with ID starting with b45c7735250ba0956fac037b363c75b63005982d5d5ca5bbae077901302d6d75 not found: ID does not exist" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.854179 4809 scope.go:117] "RemoveContainer" containerID="0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba" Dec 05 12:17:33 crc kubenswrapper[4809]: E1205 12:17:33.854793 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba\": container with ID starting with 0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba not found: ID does not exist" containerID="0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.854821 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba"} err="failed to get container status \"0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba\": rpc error: code = NotFound desc = could not find container \"0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba\": container with ID starting with 0df79ab25cb2ba1aad2f0bdd010d1dcd21755cd2b9f33e1ff9db5b818fc94eba not found: ID does not exist" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.854842 4809 scope.go:117] "RemoveContainer" containerID="419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f" Dec 05 12:17:33 crc kubenswrapper[4809]: E1205 12:17:33.855997 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f\": container with ID starting with 419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f not found: ID does not exist" containerID="419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f" Dec 05 12:17:33 crc kubenswrapper[4809]: I1205 12:17:33.856031 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f"} err="failed to get container status \"419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f\": rpc error: code = NotFound desc = could not find container \"419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f\": container with ID starting with 419d81d538dcdab119f19f04a86b1dafcd3e34cad88f20ab84aefea34a8e849f not found: ID does not exist" Dec 05 12:17:34 crc kubenswrapper[4809]: I1205 12:17:34.884912 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" path="/var/lib/kubelet/pods/e4ced3d5-2952-4344-89d8-4bbf707bd5f0/volumes" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.279778 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-smlch"] Dec 05 12:18:00 crc kubenswrapper[4809]: E1205 12:18:00.280762 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="registry-server" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.280781 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="registry-server" Dec 05 12:18:00 crc kubenswrapper[4809]: E1205 12:18:00.280813 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="extract-utilities" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.280822 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="extract-utilities" Dec 05 12:18:00 crc kubenswrapper[4809]: E1205 12:18:00.280838 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="extract-content" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.280847 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="extract-content" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.281061 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ced3d5-2952-4344-89d8-4bbf707bd5f0" containerName="registry-server" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.282543 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.293965 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-smlch"] Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.346516 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbqvd\" (UniqueName: \"kubernetes.io/projected/0986451a-2a68-4d91-9fc3-ff12c01e3434-kube-api-access-sbqvd\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.346567 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-utilities\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.346622 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-catalog-content\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.448336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-catalog-content\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.448430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbqvd\" (UniqueName: \"kubernetes.io/projected/0986451a-2a68-4d91-9fc3-ff12c01e3434-kube-api-access-sbqvd\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.448452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-utilities\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.448872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-utilities\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.448955 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-catalog-content\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.473765 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbqvd\" (UniqueName: \"kubernetes.io/projected/0986451a-2a68-4d91-9fc3-ff12c01e3434-kube-api-access-sbqvd\") pod \"redhat-operators-smlch\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:00 crc kubenswrapper[4809]: I1205 12:18:00.610616 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:01 crc kubenswrapper[4809]: I1205 12:18:01.068332 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-smlch"] Dec 05 12:18:02 crc kubenswrapper[4809]: I1205 12:18:02.002736 4809 generic.go:334] "Generic (PLEG): container finished" podID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerID="5b3cf0d3094ea476ce2f1fdb9ecaedbb24dda0d56f446f10da62508e311fb56b" exitCode=0 Dec 05 12:18:02 crc kubenswrapper[4809]: I1205 12:18:02.002844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlch" event={"ID":"0986451a-2a68-4d91-9fc3-ff12c01e3434","Type":"ContainerDied","Data":"5b3cf0d3094ea476ce2f1fdb9ecaedbb24dda0d56f446f10da62508e311fb56b"} Dec 05 12:18:02 crc kubenswrapper[4809]: I1205 12:18:02.003027 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlch" event={"ID":"0986451a-2a68-4d91-9fc3-ff12c01e3434","Type":"ContainerStarted","Data":"390c2176ea1503e7b2071cb2d4585b3e1a7ba4785d0ca7321d444273079974ee"} Dec 05 12:18:03 crc kubenswrapper[4809]: I1205 12:18:03.013736 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlch" event={"ID":"0986451a-2a68-4d91-9fc3-ff12c01e3434","Type":"ContainerStarted","Data":"eff733b36f96abc68578552ad72bd64da6eb382fcad246f6a160387f656678bf"} Dec 05 12:18:04 crc kubenswrapper[4809]: I1205 12:18:04.022129 4809 generic.go:334] "Generic (PLEG): container finished" podID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerID="eff733b36f96abc68578552ad72bd64da6eb382fcad246f6a160387f656678bf" exitCode=0 Dec 05 12:18:04 crc kubenswrapper[4809]: I1205 12:18:04.022197 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlch" event={"ID":"0986451a-2a68-4d91-9fc3-ff12c01e3434","Type":"ContainerDied","Data":"eff733b36f96abc68578552ad72bd64da6eb382fcad246f6a160387f656678bf"} Dec 05 12:18:05 crc kubenswrapper[4809]: I1205 12:18:05.037142 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlch" event={"ID":"0986451a-2a68-4d91-9fc3-ff12c01e3434","Type":"ContainerStarted","Data":"26edf454707a2ed87eabb1af7db1d83bdcb33edbe7edd5dea41ec68ac0ddee75"} Dec 05 12:18:05 crc kubenswrapper[4809]: I1205 12:18:05.071060 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-smlch" podStartSLOduration=2.617219506 podStartE2EDuration="5.071037965s" podCreationTimestamp="2025-12-05 12:18:00 +0000 UTC" firstStartedPulling="2025-12-05 12:18:02.004282507 +0000 UTC m=+4177.395259075" lastFinishedPulling="2025-12-05 12:18:04.458100976 +0000 UTC m=+4179.849077534" observedRunningTime="2025-12-05 12:18:05.06542408 +0000 UTC m=+4180.456400648" watchObservedRunningTime="2025-12-05 12:18:05.071037965 +0000 UTC m=+4180.462014533" Dec 05 12:18:10 crc kubenswrapper[4809]: I1205 12:18:10.612551 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:10 crc kubenswrapper[4809]: I1205 12:18:10.613125 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:10 crc kubenswrapper[4809]: I1205 12:18:10.660435 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:11 crc kubenswrapper[4809]: I1205 12:18:11.117995 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:11 crc kubenswrapper[4809]: I1205 12:18:11.165866 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-smlch"] Dec 05 12:18:13 crc kubenswrapper[4809]: I1205 12:18:13.089688 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-smlch" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="registry-server" containerID="cri-o://26edf454707a2ed87eabb1af7db1d83bdcb33edbe7edd5dea41ec68ac0ddee75" gracePeriod=2 Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.119954 4809 generic.go:334] "Generic (PLEG): container finished" podID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerID="26edf454707a2ed87eabb1af7db1d83bdcb33edbe7edd5dea41ec68ac0ddee75" exitCode=0 Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.120001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlch" event={"ID":"0986451a-2a68-4d91-9fc3-ff12c01e3434","Type":"ContainerDied","Data":"26edf454707a2ed87eabb1af7db1d83bdcb33edbe7edd5dea41ec68ac0ddee75"} Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.442272 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.498681 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbqvd\" (UniqueName: \"kubernetes.io/projected/0986451a-2a68-4d91-9fc3-ff12c01e3434-kube-api-access-sbqvd\") pod \"0986451a-2a68-4d91-9fc3-ff12c01e3434\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.498736 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-utilities\") pod \"0986451a-2a68-4d91-9fc3-ff12c01e3434\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.498765 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-catalog-content\") pod \"0986451a-2a68-4d91-9fc3-ff12c01e3434\" (UID: \"0986451a-2a68-4d91-9fc3-ff12c01e3434\") " Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.500851 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-utilities" (OuterVolumeSpecName: "utilities") pod "0986451a-2a68-4d91-9fc3-ff12c01e3434" (UID: "0986451a-2a68-4d91-9fc3-ff12c01e3434"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.513617 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0986451a-2a68-4d91-9fc3-ff12c01e3434-kube-api-access-sbqvd" (OuterVolumeSpecName: "kube-api-access-sbqvd") pod "0986451a-2a68-4d91-9fc3-ff12c01e3434" (UID: "0986451a-2a68-4d91-9fc3-ff12c01e3434"). InnerVolumeSpecName "kube-api-access-sbqvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.600517 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbqvd\" (UniqueName: \"kubernetes.io/projected/0986451a-2a68-4d91-9fc3-ff12c01e3434-kube-api-access-sbqvd\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.600549 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.605288 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0986451a-2a68-4d91-9fc3-ff12c01e3434" (UID: "0986451a-2a68-4d91-9fc3-ff12c01e3434"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:18:16 crc kubenswrapper[4809]: I1205 12:18:16.702292 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0986451a-2a68-4d91-9fc3-ff12c01e3434-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:18:17 crc kubenswrapper[4809]: I1205 12:18:17.128750 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlch" event={"ID":"0986451a-2a68-4d91-9fc3-ff12c01e3434","Type":"ContainerDied","Data":"390c2176ea1503e7b2071cb2d4585b3e1a7ba4785d0ca7321d444273079974ee"} Dec 05 12:18:17 crc kubenswrapper[4809]: I1205 12:18:17.128816 4809 scope.go:117] "RemoveContainer" containerID="26edf454707a2ed87eabb1af7db1d83bdcb33edbe7edd5dea41ec68ac0ddee75" Dec 05 12:18:17 crc kubenswrapper[4809]: I1205 12:18:17.128832 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlch" Dec 05 12:18:17 crc kubenswrapper[4809]: I1205 12:18:17.150466 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-smlch"] Dec 05 12:18:17 crc kubenswrapper[4809]: I1205 12:18:17.156788 4809 scope.go:117] "RemoveContainer" containerID="eff733b36f96abc68578552ad72bd64da6eb382fcad246f6a160387f656678bf" Dec 05 12:18:17 crc kubenswrapper[4809]: I1205 12:18:17.157831 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-smlch"] Dec 05 12:18:17 crc kubenswrapper[4809]: I1205 12:18:17.176583 4809 scope.go:117] "RemoveContainer" containerID="5b3cf0d3094ea476ce2f1fdb9ecaedbb24dda0d56f446f10da62508e311fb56b" Dec 05 12:18:18 crc kubenswrapper[4809]: I1205 12:18:18.882529 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" path="/var/lib/kubelet/pods/0986451a-2a68-4d91-9fc3-ff12c01e3434/volumes" Dec 05 12:19:14 crc kubenswrapper[4809]: I1205 12:19:14.046869 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:19:14 crc kubenswrapper[4809]: I1205 12:19:14.049230 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:19:44 crc kubenswrapper[4809]: I1205 12:19:44.046468 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:19:44 crc kubenswrapper[4809]: I1205 12:19:44.046958 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.046914 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.047492 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.047541 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.048280 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.048360 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" gracePeriod=600 Dec 05 12:20:14 crc kubenswrapper[4809]: E1205 12:20:14.173655 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.370994 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" exitCode=0 Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.371052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8"} Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.371102 4809 scope.go:117] "RemoveContainer" containerID="d9a9dd09031827650d11d1b983fac8b8ba6eeb9c83c05f84175ab691223ff77f" Dec 05 12:20:14 crc kubenswrapper[4809]: I1205 12:20:14.372841 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:20:14 crc kubenswrapper[4809]: E1205 12:20:14.373280 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:20:26 crc kubenswrapper[4809]: I1205 12:20:26.871944 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:20:26 crc kubenswrapper[4809]: E1205 12:20:26.872537 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:20:41 crc kubenswrapper[4809]: I1205 12:20:41.872900 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:20:41 crc kubenswrapper[4809]: E1205 12:20:41.873825 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:20:52 crc kubenswrapper[4809]: I1205 12:20:52.872964 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:20:52 crc kubenswrapper[4809]: E1205 12:20:52.873714 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:21:04 crc kubenswrapper[4809]: I1205 12:21:04.876892 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:21:04 crc kubenswrapper[4809]: E1205 12:21:04.877688 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:21:17 crc kubenswrapper[4809]: I1205 12:21:17.873575 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:21:17 crc kubenswrapper[4809]: E1205 12:21:17.874539 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:21:28 crc kubenswrapper[4809]: I1205 12:21:28.873681 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:21:28 crc kubenswrapper[4809]: E1205 12:21:28.874615 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:21:42 crc kubenswrapper[4809]: I1205 12:21:42.872114 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:21:42 crc kubenswrapper[4809]: E1205 12:21:42.874024 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:21:54 crc kubenswrapper[4809]: I1205 12:21:54.876668 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:21:54 crc kubenswrapper[4809]: E1205 12:21:54.877605 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:22:07 crc kubenswrapper[4809]: I1205 12:22:07.872318 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:22:07 crc kubenswrapper[4809]: E1205 12:22:07.873494 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:22:21 crc kubenswrapper[4809]: I1205 12:22:21.873227 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:22:21 crc kubenswrapper[4809]: E1205 12:22:21.874738 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:22:33 crc kubenswrapper[4809]: I1205 12:22:33.872578 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:22:33 crc kubenswrapper[4809]: E1205 12:22:33.873681 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:22:46 crc kubenswrapper[4809]: I1205 12:22:46.872081 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:22:46 crc kubenswrapper[4809]: E1205 12:22:46.872942 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:22:59 crc kubenswrapper[4809]: I1205 12:22:59.874187 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:22:59 crc kubenswrapper[4809]: E1205 12:22:59.875030 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:23:10 crc kubenswrapper[4809]: I1205 12:23:10.872521 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:23:10 crc kubenswrapper[4809]: E1205 12:23:10.873121 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:23:25 crc kubenswrapper[4809]: I1205 12:23:25.873037 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:23:25 crc kubenswrapper[4809]: E1205 12:23:25.873875 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:23:36 crc kubenswrapper[4809]: I1205 12:23:36.872595 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:23:36 crc kubenswrapper[4809]: E1205 12:23:36.873489 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:23:48 crc kubenswrapper[4809]: I1205 12:23:48.872131 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:23:48 crc kubenswrapper[4809]: E1205 12:23:48.872862 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:24:00 crc kubenswrapper[4809]: I1205 12:24:00.871924 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:24:00 crc kubenswrapper[4809]: E1205 12:24:00.872672 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:24:11 crc kubenswrapper[4809]: I1205 12:24:11.871929 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:24:11 crc kubenswrapper[4809]: E1205 12:24:11.872734 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:24:24 crc kubenswrapper[4809]: I1205 12:24:24.877331 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:24:24 crc kubenswrapper[4809]: E1205 12:24:24.880052 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.574995 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-xxhbm"] Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.582137 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-xxhbm"] Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.695520 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-sj6fw"] Dec 05 12:24:30 crc kubenswrapper[4809]: E1205 12:24:30.695891 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="registry-server" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.695906 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="registry-server" Dec 05 12:24:30 crc kubenswrapper[4809]: E1205 12:24:30.695919 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="extract-utilities" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.695926 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="extract-utilities" Dec 05 12:24:30 crc kubenswrapper[4809]: E1205 12:24:30.695948 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="extract-content" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.695955 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="extract-content" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.696119 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0986451a-2a68-4d91-9fc3-ff12c01e3434" containerName="registry-server" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.696737 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.699479 4809 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-lngh5" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.699899 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.700150 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.704870 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.704941 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-sj6fw"] Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.720856 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9e00195d-9c4c-4232-a850-2bd4eac47871-node-mnt\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.720919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9e00195d-9c4c-4232-a850-2bd4eac47871-crc-storage\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.720952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sjh8\" (UniqueName: \"kubernetes.io/projected/9e00195d-9c4c-4232-a850-2bd4eac47871-kube-api-access-7sjh8\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.822451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9e00195d-9c4c-4232-a850-2bd4eac47871-node-mnt\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.823595 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9e00195d-9c4c-4232-a850-2bd4eac47871-crc-storage\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.823731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sjh8\" (UniqueName: \"kubernetes.io/projected/9e00195d-9c4c-4232-a850-2bd4eac47871-kube-api-access-7sjh8\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.823831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9e00195d-9c4c-4232-a850-2bd4eac47871-node-mnt\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.824417 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9e00195d-9c4c-4232-a850-2bd4eac47871-crc-storage\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.849594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sjh8\" (UniqueName: \"kubernetes.io/projected/9e00195d-9c4c-4232-a850-2bd4eac47871-kube-api-access-7sjh8\") pod \"crc-storage-crc-sj6fw\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:30 crc kubenswrapper[4809]: I1205 12:24:30.882780 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f4be64e-7da9-41d2-95b6-bf38e7cc254e" path="/var/lib/kubelet/pods/6f4be64e-7da9-41d2-95b6-bf38e7cc254e/volumes" Dec 05 12:24:31 crc kubenswrapper[4809]: I1205 12:24:31.021891 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:31 crc kubenswrapper[4809]: I1205 12:24:31.414508 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-sj6fw"] Dec 05 12:24:31 crc kubenswrapper[4809]: I1205 12:24:31.422951 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:24:31 crc kubenswrapper[4809]: I1205 12:24:31.557031 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-sj6fw" event={"ID":"9e00195d-9c4c-4232-a850-2bd4eac47871","Type":"ContainerStarted","Data":"a1b5b9551494ad0dad1cf42927ba24fa9f610d20f32020b9e9014f05c073d7fa"} Dec 05 12:24:32 crc kubenswrapper[4809]: I1205 12:24:32.565703 4809 generic.go:334] "Generic (PLEG): container finished" podID="9e00195d-9c4c-4232-a850-2bd4eac47871" containerID="148e51844f6d4b3633a336f8810585ea8ac01c904ef4e6f692a9891fc72e3b75" exitCode=0 Dec 05 12:24:32 crc kubenswrapper[4809]: I1205 12:24:32.565784 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-sj6fw" event={"ID":"9e00195d-9c4c-4232-a850-2bd4eac47871","Type":"ContainerDied","Data":"148e51844f6d4b3633a336f8810585ea8ac01c904ef4e6f692a9891fc72e3b75"} Dec 05 12:24:33 crc kubenswrapper[4809]: I1205 12:24:33.886533 4809 scope.go:117] "RemoveContainer" containerID="eb7bddf66b3cfad78ed11ced857ef43a6d395bcb8aed6b3ee9b02775b51d267b" Dec 05 12:24:33 crc kubenswrapper[4809]: I1205 12:24:33.942899 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.068577 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9e00195d-9c4c-4232-a850-2bd4eac47871-crc-storage\") pod \"9e00195d-9c4c-4232-a850-2bd4eac47871\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.068712 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sjh8\" (UniqueName: \"kubernetes.io/projected/9e00195d-9c4c-4232-a850-2bd4eac47871-kube-api-access-7sjh8\") pod \"9e00195d-9c4c-4232-a850-2bd4eac47871\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.068839 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9e00195d-9c4c-4232-a850-2bd4eac47871-node-mnt\") pod \"9e00195d-9c4c-4232-a850-2bd4eac47871\" (UID: \"9e00195d-9c4c-4232-a850-2bd4eac47871\") " Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.069824 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e00195d-9c4c-4232-a850-2bd4eac47871-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "9e00195d-9c4c-4232-a850-2bd4eac47871" (UID: "9e00195d-9c4c-4232-a850-2bd4eac47871"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.073570 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e00195d-9c4c-4232-a850-2bd4eac47871-kube-api-access-7sjh8" (OuterVolumeSpecName: "kube-api-access-7sjh8") pod "9e00195d-9c4c-4232-a850-2bd4eac47871" (UID: "9e00195d-9c4c-4232-a850-2bd4eac47871"). InnerVolumeSpecName "kube-api-access-7sjh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.086909 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e00195d-9c4c-4232-a850-2bd4eac47871-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "9e00195d-9c4c-4232-a850-2bd4eac47871" (UID: "9e00195d-9c4c-4232-a850-2bd4eac47871"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.170614 4809 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9e00195d-9c4c-4232-a850-2bd4eac47871-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.170671 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sjh8\" (UniqueName: \"kubernetes.io/projected/9e00195d-9c4c-4232-a850-2bd4eac47871-kube-api-access-7sjh8\") on node \"crc\" DevicePath \"\"" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.170680 4809 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9e00195d-9c4c-4232-a850-2bd4eac47871-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.585578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-sj6fw" event={"ID":"9e00195d-9c4c-4232-a850-2bd4eac47871","Type":"ContainerDied","Data":"a1b5b9551494ad0dad1cf42927ba24fa9f610d20f32020b9e9014f05c073d7fa"} Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.585679 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1b5b9551494ad0dad1cf42927ba24fa9f610d20f32020b9e9014f05c073d7fa" Dec 05 12:24:34 crc kubenswrapper[4809]: I1205 12:24:34.585759 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-sj6fw" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.048685 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-sj6fw"] Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.053901 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-sj6fw"] Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.194986 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-drbbt"] Dec 05 12:24:36 crc kubenswrapper[4809]: E1205 12:24:36.195363 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e00195d-9c4c-4232-a850-2bd4eac47871" containerName="storage" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.195385 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e00195d-9c4c-4232-a850-2bd4eac47871" containerName="storage" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.195553 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e00195d-9c4c-4232-a850-2bd4eac47871" containerName="storage" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.196156 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.203802 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.204023 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.204169 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.204406 4809 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-lngh5" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.204930 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-drbbt"] Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.297849 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976ww\" (UniqueName: \"kubernetes.io/projected/edaa40fb-393d-446d-8577-98eb17bfeebf-kube-api-access-976ww\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.297923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edaa40fb-393d-446d-8577-98eb17bfeebf-crc-storage\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.297950 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edaa40fb-393d-446d-8577-98eb17bfeebf-node-mnt\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.399739 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976ww\" (UniqueName: \"kubernetes.io/projected/edaa40fb-393d-446d-8577-98eb17bfeebf-kube-api-access-976ww\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.399781 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edaa40fb-393d-446d-8577-98eb17bfeebf-crc-storage\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.399800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edaa40fb-393d-446d-8577-98eb17bfeebf-node-mnt\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.400062 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edaa40fb-393d-446d-8577-98eb17bfeebf-node-mnt\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.400872 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edaa40fb-393d-446d-8577-98eb17bfeebf-crc-storage\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.421038 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976ww\" (UniqueName: \"kubernetes.io/projected/edaa40fb-393d-446d-8577-98eb17bfeebf-kube-api-access-976ww\") pod \"crc-storage-crc-drbbt\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.513465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.880558 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e00195d-9c4c-4232-a850-2bd4eac47871" path="/var/lib/kubelet/pods/9e00195d-9c4c-4232-a850-2bd4eac47871/volumes" Dec 05 12:24:36 crc kubenswrapper[4809]: I1205 12:24:36.926142 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-drbbt"] Dec 05 12:24:37 crc kubenswrapper[4809]: I1205 12:24:37.608410 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-drbbt" event={"ID":"edaa40fb-393d-446d-8577-98eb17bfeebf","Type":"ContainerStarted","Data":"537abd87daf50e90ade07d276e914e742f0bde31e6b6fed14c8b4d0644786a03"} Dec 05 12:24:37 crc kubenswrapper[4809]: I1205 12:24:37.873122 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:24:37 crc kubenswrapper[4809]: E1205 12:24:37.873580 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:24:38 crc kubenswrapper[4809]: I1205 12:24:38.615417 4809 generic.go:334] "Generic (PLEG): container finished" podID="edaa40fb-393d-446d-8577-98eb17bfeebf" containerID="629518040da378fcf60ecde52a675ad9b96ab754c1c0fe9c5f11e25c99ea3eb8" exitCode=0 Dec 05 12:24:38 crc kubenswrapper[4809]: I1205 12:24:38.615461 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-drbbt" event={"ID":"edaa40fb-393d-446d-8577-98eb17bfeebf","Type":"ContainerDied","Data":"629518040da378fcf60ecde52a675ad9b96ab754c1c0fe9c5f11e25c99ea3eb8"} Dec 05 12:24:39 crc kubenswrapper[4809]: I1205 12:24:39.952579 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.050978 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edaa40fb-393d-446d-8577-98eb17bfeebf-node-mnt\") pod \"edaa40fb-393d-446d-8577-98eb17bfeebf\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.051063 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edaa40fb-393d-446d-8577-98eb17bfeebf-crc-storage\") pod \"edaa40fb-393d-446d-8577-98eb17bfeebf\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.051115 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/edaa40fb-393d-446d-8577-98eb17bfeebf-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "edaa40fb-393d-446d-8577-98eb17bfeebf" (UID: "edaa40fb-393d-446d-8577-98eb17bfeebf"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.051256 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976ww\" (UniqueName: \"kubernetes.io/projected/edaa40fb-393d-446d-8577-98eb17bfeebf-kube-api-access-976ww\") pod \"edaa40fb-393d-446d-8577-98eb17bfeebf\" (UID: \"edaa40fb-393d-446d-8577-98eb17bfeebf\") " Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.051869 4809 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edaa40fb-393d-446d-8577-98eb17bfeebf-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.056108 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edaa40fb-393d-446d-8577-98eb17bfeebf-kube-api-access-976ww" (OuterVolumeSpecName: "kube-api-access-976ww") pod "edaa40fb-393d-446d-8577-98eb17bfeebf" (UID: "edaa40fb-393d-446d-8577-98eb17bfeebf"). InnerVolumeSpecName "kube-api-access-976ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.067543 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edaa40fb-393d-446d-8577-98eb17bfeebf-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "edaa40fb-393d-446d-8577-98eb17bfeebf" (UID: "edaa40fb-393d-446d-8577-98eb17bfeebf"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.153686 4809 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edaa40fb-393d-446d-8577-98eb17bfeebf-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.154048 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976ww\" (UniqueName: \"kubernetes.io/projected/edaa40fb-393d-446d-8577-98eb17bfeebf-kube-api-access-976ww\") on node \"crc\" DevicePath \"\"" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.633205 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-drbbt" event={"ID":"edaa40fb-393d-446d-8577-98eb17bfeebf","Type":"ContainerDied","Data":"537abd87daf50e90ade07d276e914e742f0bde31e6b6fed14c8b4d0644786a03"} Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.633251 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="537abd87daf50e90ade07d276e914e742f0bde31e6b6fed14c8b4d0644786a03" Dec 05 12:24:40 crc kubenswrapper[4809]: I1205 12:24:40.633360 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-drbbt" Dec 05 12:24:49 crc kubenswrapper[4809]: I1205 12:24:49.873181 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:24:49 crc kubenswrapper[4809]: E1205 12:24:49.874734 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:25:00 crc kubenswrapper[4809]: I1205 12:25:00.872937 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:25:00 crc kubenswrapper[4809]: E1205 12:25:00.873762 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:25:15 crc kubenswrapper[4809]: I1205 12:25:15.873074 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:25:16 crc kubenswrapper[4809]: I1205 12:25:16.934308 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"90fcfc02adb974dcd81059abed9da13265a950cbccc4646c761fd364738d11bd"} Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.787238 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwlqn"] Dec 05 12:25:24 crc kubenswrapper[4809]: E1205 12:25:24.788135 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edaa40fb-393d-446d-8577-98eb17bfeebf" containerName="storage" Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.788148 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="edaa40fb-393d-446d-8577-98eb17bfeebf" containerName="storage" Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.788317 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="edaa40fb-393d-446d-8577-98eb17bfeebf" containerName="storage" Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.789315 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.793897 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwlqn"] Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.959411 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqwtg\" (UniqueName: \"kubernetes.io/projected/fe8e3454-568a-4453-b0af-fb4427152cd4-kube-api-access-dqwtg\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.959472 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-catalog-content\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:24 crc kubenswrapper[4809]: I1205 12:25:24.959620 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-utilities\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.060921 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-utilities\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.060988 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqwtg\" (UniqueName: \"kubernetes.io/projected/fe8e3454-568a-4453-b0af-fb4427152cd4-kube-api-access-dqwtg\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.061011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-catalog-content\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.062415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-utilities\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.062456 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-catalog-content\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.101528 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqwtg\" (UniqueName: \"kubernetes.io/projected/fe8e3454-568a-4453-b0af-fb4427152cd4-kube-api-access-dqwtg\") pod \"community-operators-nwlqn\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.118133 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:25 crc kubenswrapper[4809]: I1205 12:25:25.610748 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwlqn"] Dec 05 12:25:26 crc kubenswrapper[4809]: I1205 12:25:26.008758 4809 generic.go:334] "Generic (PLEG): container finished" podID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerID="7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36" exitCode=0 Dec 05 12:25:26 crc kubenswrapper[4809]: I1205 12:25:26.008817 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwlqn" event={"ID":"fe8e3454-568a-4453-b0af-fb4427152cd4","Type":"ContainerDied","Data":"7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36"} Dec 05 12:25:26 crc kubenswrapper[4809]: I1205 12:25:26.009260 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwlqn" event={"ID":"fe8e3454-568a-4453-b0af-fb4427152cd4","Type":"ContainerStarted","Data":"90cb503342a08d1eb31d75d6d62f0c45a784b7bd9ade192018eb3b997b6386b2"} Dec 05 12:25:27 crc kubenswrapper[4809]: I1205 12:25:27.017960 4809 generic.go:334] "Generic (PLEG): container finished" podID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerID="e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb" exitCode=0 Dec 05 12:25:27 crc kubenswrapper[4809]: I1205 12:25:27.018170 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwlqn" event={"ID":"fe8e3454-568a-4453-b0af-fb4427152cd4","Type":"ContainerDied","Data":"e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb"} Dec 05 12:25:28 crc kubenswrapper[4809]: I1205 12:25:28.036003 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwlqn" event={"ID":"fe8e3454-568a-4453-b0af-fb4427152cd4","Type":"ContainerStarted","Data":"425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc"} Dec 05 12:25:28 crc kubenswrapper[4809]: I1205 12:25:28.060256 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwlqn" podStartSLOduration=2.635202507 podStartE2EDuration="4.060235535s" podCreationTimestamp="2025-12-05 12:25:24 +0000 UTC" firstStartedPulling="2025-12-05 12:25:26.011014369 +0000 UTC m=+4621.401990947" lastFinishedPulling="2025-12-05 12:25:27.436047407 +0000 UTC m=+4622.827023975" observedRunningTime="2025-12-05 12:25:28.051756516 +0000 UTC m=+4623.442733094" watchObservedRunningTime="2025-12-05 12:25:28.060235535 +0000 UTC m=+4623.451212093" Dec 05 12:25:35 crc kubenswrapper[4809]: I1205 12:25:35.118778 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:35 crc kubenswrapper[4809]: I1205 12:25:35.119390 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:35 crc kubenswrapper[4809]: I1205 12:25:35.419153 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:36 crc kubenswrapper[4809]: I1205 12:25:36.159528 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:37 crc kubenswrapper[4809]: I1205 12:25:37.362944 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwlqn"] Dec 05 12:25:38 crc kubenswrapper[4809]: I1205 12:25:38.127714 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwlqn" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="registry-server" containerID="cri-o://425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc" gracePeriod=2 Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.063331 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.139970 4809 generic.go:334] "Generic (PLEG): container finished" podID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerID="425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc" exitCode=0 Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.140019 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwlqn" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.140020 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwlqn" event={"ID":"fe8e3454-568a-4453-b0af-fb4427152cd4","Type":"ContainerDied","Data":"425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc"} Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.140137 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwlqn" event={"ID":"fe8e3454-568a-4453-b0af-fb4427152cd4","Type":"ContainerDied","Data":"90cb503342a08d1eb31d75d6d62f0c45a784b7bd9ade192018eb3b997b6386b2"} Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.140162 4809 scope.go:117] "RemoveContainer" containerID="425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.160323 4809 scope.go:117] "RemoveContainer" containerID="e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.185395 4809 scope.go:117] "RemoveContainer" containerID="7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.186322 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqwtg\" (UniqueName: \"kubernetes.io/projected/fe8e3454-568a-4453-b0af-fb4427152cd4-kube-api-access-dqwtg\") pod \"fe8e3454-568a-4453-b0af-fb4427152cd4\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.186516 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-catalog-content\") pod \"fe8e3454-568a-4453-b0af-fb4427152cd4\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.186553 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-utilities\") pod \"fe8e3454-568a-4453-b0af-fb4427152cd4\" (UID: \"fe8e3454-568a-4453-b0af-fb4427152cd4\") " Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.188233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-utilities" (OuterVolumeSpecName: "utilities") pod "fe8e3454-568a-4453-b0af-fb4427152cd4" (UID: "fe8e3454-568a-4453-b0af-fb4427152cd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.192818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe8e3454-568a-4453-b0af-fb4427152cd4-kube-api-access-dqwtg" (OuterVolumeSpecName: "kube-api-access-dqwtg") pod "fe8e3454-568a-4453-b0af-fb4427152cd4" (UID: "fe8e3454-568a-4453-b0af-fb4427152cd4"). InnerVolumeSpecName "kube-api-access-dqwtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.233792 4809 scope.go:117] "RemoveContainer" containerID="425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc" Dec 05 12:25:39 crc kubenswrapper[4809]: E1205 12:25:39.234343 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc\": container with ID starting with 425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc not found: ID does not exist" containerID="425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.234402 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc"} err="failed to get container status \"425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc\": rpc error: code = NotFound desc = could not find container \"425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc\": container with ID starting with 425436b72288ed834955c1dc3077400468bfbbc928d0b6a9136933122a2072fc not found: ID does not exist" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.234437 4809 scope.go:117] "RemoveContainer" containerID="e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb" Dec 05 12:25:39 crc kubenswrapper[4809]: E1205 12:25:39.234982 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb\": container with ID starting with e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb not found: ID does not exist" containerID="e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.235049 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb"} err="failed to get container status \"e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb\": rpc error: code = NotFound desc = could not find container \"e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb\": container with ID starting with e5570cbd5ee576c3260183802231ae52e35d4e38e2b29444faa13a0322671aeb not found: ID does not exist" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.235086 4809 scope.go:117] "RemoveContainer" containerID="7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36" Dec 05 12:25:39 crc kubenswrapper[4809]: E1205 12:25:39.235552 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36\": container with ID starting with 7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36 not found: ID does not exist" containerID="7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.235589 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36"} err="failed to get container status \"7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36\": rpc error: code = NotFound desc = could not find container \"7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36\": container with ID starting with 7b463a8af00ab0ffdd2aaf6fd8617336ebce56b817833800df867299abe48f36 not found: ID does not exist" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.238965 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe8e3454-568a-4453-b0af-fb4427152cd4" (UID: "fe8e3454-568a-4453-b0af-fb4427152cd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.288980 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqwtg\" (UniqueName: \"kubernetes.io/projected/fe8e3454-568a-4453-b0af-fb4427152cd4-kube-api-access-dqwtg\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.289010 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.289021 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8e3454-568a-4453-b0af-fb4427152cd4-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.470777 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwlqn"] Dec 05 12:25:39 crc kubenswrapper[4809]: I1205 12:25:39.476687 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwlqn"] Dec 05 12:25:40 crc kubenswrapper[4809]: I1205 12:25:40.883256 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" path="/var/lib/kubelet/pods/fe8e3454-568a-4453-b0af-fb4427152cd4/volumes" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.575450 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-ptwr6"] Dec 05 12:27:43 crc kubenswrapper[4809]: E1205 12:27:43.577669 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="registry-server" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.577771 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="registry-server" Dec 05 12:27:43 crc kubenswrapper[4809]: E1205 12:27:43.577849 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="extract-utilities" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.577910 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="extract-utilities" Dec 05 12:27:43 crc kubenswrapper[4809]: E1205 12:27:43.577969 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="extract-content" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.578025 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="extract-content" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.578198 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe8e3454-568a-4453-b0af-fb4427152cd4" containerName="registry-server" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.579046 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.582053 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.582339 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-2trp2" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.583069 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.583726 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.588044 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.604173 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-ptwr6"] Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.693457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.693594 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4vvc\" (UniqueName: \"kubernetes.io/projected/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-kube-api-access-q4vvc\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.693621 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-config\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.794992 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.795156 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4vvc\" (UniqueName: \"kubernetes.io/projected/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-kube-api-access-q4vvc\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.795189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-config\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.796130 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.796201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-config\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.834494 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4vvc\" (UniqueName: \"kubernetes.io/projected/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-kube-api-access-q4vvc\") pod \"dnsmasq-dns-5d7b5456f5-ptwr6\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.841024 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-9gdsv"] Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.842513 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.855339 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-9gdsv"] Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.894902 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.896527 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.896720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzvsw\" (UniqueName: \"kubernetes.io/projected/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-kube-api-access-xzvsw\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:43 crc kubenswrapper[4809]: I1205 12:27:43.896900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-config\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.013977 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-config\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.014055 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.014089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzvsw\" (UniqueName: \"kubernetes.io/projected/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-kube-api-access-xzvsw\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.015508 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-config\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.016246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.047683 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.047768 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.049890 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzvsw\" (UniqueName: \"kubernetes.io/projected/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-kube-api-access-xzvsw\") pod \"dnsmasq-dns-98ddfc8f-9gdsv\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.185870 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-ptwr6"] Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.188850 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.327268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" event={"ID":"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e","Type":"ContainerStarted","Data":"7561176eb2179661388f6c826d05144ae550c9289ef709123648a7212dcc40a3"} Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.624404 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-9gdsv"] Dec 05 12:27:44 crc kubenswrapper[4809]: W1205 12:27:44.627089 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfde5555e_3d40_4b92_ab47_f9b2043ddb1b.slice/crio-970d59866285e5608cb3f78b8d3ba8d195fbecc16987f81e5b8f41a11859ab14 WatchSource:0}: Error finding container 970d59866285e5608cb3f78b8d3ba8d195fbecc16987f81e5b8f41a11859ab14: Status 404 returned error can't find the container with id 970d59866285e5608cb3f78b8d3ba8d195fbecc16987f81e5b8f41a11859ab14 Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.632612 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.633786 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.635412 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.636484 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7pfpn" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.636619 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.636837 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.636967 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.658941 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.824922 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f15947f-ce8c-42b9-8221-aae51566ebf8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.824971 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.825014 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.825041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.825066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.825083 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2klf\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-kube-api-access-v2klf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.825108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f15947f-ce8c-42b9-8221-aae51566ebf8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.825158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.825178 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.944589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2klf\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-kube-api-access-v2klf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946258 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f15947f-ce8c-42b9-8221-aae51566ebf8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946655 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f15947f-ce8c-42b9-8221-aae51566ebf8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946735 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.946836 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.947192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.945990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.949484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.951387 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.952468 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f15947f-ce8c-42b9-8221-aae51566ebf8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.953361 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.955360 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f15947f-ce8c-42b9-8221-aae51566ebf8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.960048 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.960101 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e22c51975a412c7ed096050c61fe20a6ae4ba86f0578c4586a4dfb3a62e17ae5/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.972621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2klf\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-kube-api-access-v2klf\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.991643 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:27:44 crc kubenswrapper[4809]: I1205 12:27:44.992928 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:44.998045 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:44.998233 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-t9ch6" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:44.998351 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:44.998457 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:44.998586 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.015907 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.048878 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " pod="openstack/rabbitmq-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148678 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dgk6\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-kube-api-access-2dgk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148718 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148757 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148781 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148828 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148844 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148861 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.148911 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.250135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.250188 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.250217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dgk6\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-kube-api-access-2dgk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251070 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251097 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251156 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251375 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251870 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.251939 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.250998 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.252870 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.255520 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.255546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.256064 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.260804 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.260831 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cef7a9c2f28e824d64b9cf8b9bdb2eb2a26949218d41e5db25547eafe60575a4/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.282451 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dgk6\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-kube-api-access-2dgk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.296835 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.351828 4809 generic.go:334] "Generic (PLEG): container finished" podID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerID="dd45f23e4bc76c0e1e489627c0bc7e8d3b39515b78dcd5fae57c662667a8dafb" exitCode=0 Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.351893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" event={"ID":"fde5555e-3d40-4b92-ab47-f9b2043ddb1b","Type":"ContainerDied","Data":"dd45f23e4bc76c0e1e489627c0bc7e8d3b39515b78dcd5fae57c662667a8dafb"} Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.351927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" event={"ID":"fde5555e-3d40-4b92-ab47-f9b2043ddb1b","Type":"ContainerStarted","Data":"970d59866285e5608cb3f78b8d3ba8d195fbecc16987f81e5b8f41a11859ab14"} Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.358579 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.362869 4809 generic.go:334] "Generic (PLEG): container finished" podID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerID="d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6" exitCode=0 Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.362902 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" event={"ID":"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e","Type":"ContainerDied","Data":"d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6"} Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.516825 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:27:45 crc kubenswrapper[4809]: I1205 12:27:45.820947 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:27:45 crc kubenswrapper[4809]: W1205 12:27:45.828444 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee55ec3a_b3e8_4d18_bac6_4dc66ddabb87.slice/crio-8356bcb35137dfec5fed33e64c3b88654e34459708782f7e4ebaf3469200a585 WatchSource:0}: Error finding container 8356bcb35137dfec5fed33e64c3b88654e34459708782f7e4ebaf3469200a585: Status 404 returned error can't find the container with id 8356bcb35137dfec5fed33e64c3b88654e34459708782f7e4ebaf3469200a585 Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.023196 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.024900 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.027645 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.027708 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.028015 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-x4d86" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.028055 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.040004 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.044990 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.168846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.168888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpc8t\" (UniqueName: \"kubernetes.io/projected/5976c836-3783-481b-8dcb-8b5725a16e4c-kube-api-access-rpc8t\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.168927 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.168952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5976c836-3783-481b-8dcb-8b5725a16e4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.168999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-85ed96f7-e326-4ead-9d02-53d57789a556\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-85ed96f7-e326-4ead-9d02-53d57789a556\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.169026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.169041 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5976c836-3783-481b-8dcb-8b5725a16e4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.169248 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5976c836-3783-481b-8dcb-8b5725a16e4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270618 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-85ed96f7-e326-4ead-9d02-53d57789a556\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-85ed96f7-e326-4ead-9d02-53d57789a556\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270701 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270726 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5976c836-3783-481b-8dcb-8b5725a16e4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270784 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5976c836-3783-481b-8dcb-8b5725a16e4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpc8t\" (UniqueName: \"kubernetes.io/projected/5976c836-3783-481b-8dcb-8b5725a16e4c-kube-api-access-rpc8t\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270873 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.270893 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5976c836-3783-481b-8dcb-8b5725a16e4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.271860 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5976c836-3783-481b-8dcb-8b5725a16e4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.272042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.272234 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.273406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5976c836-3783-481b-8dcb-8b5725a16e4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.275409 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.275459 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-85ed96f7-e326-4ead-9d02-53d57789a556\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-85ed96f7-e326-4ead-9d02-53d57789a556\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dcf794ac9b12769b687505049948e56655d04f459a9fc1ab95b657a8533dde83/globalmount\"" pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.378587 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5976c836-3783-481b-8dcb-8b5725a16e4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.380698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5976c836-3783-481b-8dcb-8b5725a16e4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.381589 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpc8t\" (UniqueName: \"kubernetes.io/projected/5976c836-3783-481b-8dcb-8b5725a16e4c-kube-api-access-rpc8t\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.383376 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" event={"ID":"fde5555e-3d40-4b92-ab47-f9b2043ddb1b","Type":"ContainerStarted","Data":"adc6b1eedf990527e5509a77d0678ce0582266094a8fb061799998318c6d30f0"} Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.383544 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.386185 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f15947f-ce8c-42b9-8221-aae51566ebf8","Type":"ContainerStarted","Data":"79b522386ce8bfed4c2549b97775bc1ee3f80b7a54d1d97d2a29f6a0d7e03e21"} Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.389323 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" event={"ID":"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e","Type":"ContainerStarted","Data":"738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1"} Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.390092 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.391725 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87","Type":"ContainerStarted","Data":"8356bcb35137dfec5fed33e64c3b88654e34459708782f7e4ebaf3469200a585"} Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.408929 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" podStartSLOduration=3.4089066470000002 podStartE2EDuration="3.408906647s" podCreationTimestamp="2025-12-05 12:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:27:46.400133131 +0000 UTC m=+4761.791109689" watchObservedRunningTime="2025-12-05 12:27:46.408906647 +0000 UTC m=+4761.799883215" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.416536 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" podStartSLOduration=3.4165146220000002 podStartE2EDuration="3.416514622s" podCreationTimestamp="2025-12-05 12:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:27:46.415506395 +0000 UTC m=+4761.806482983" watchObservedRunningTime="2025-12-05 12:27:46.416514622 +0000 UTC m=+4761.807491180" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.497985 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-85ed96f7-e326-4ead-9d02-53d57789a556\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-85ed96f7-e326-4ead-9d02-53d57789a556\") pod \"openstack-galera-0\" (UID: \"5976c836-3783-481b-8dcb-8b5725a16e4c\") " pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.539130 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.540350 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.543023 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-c4lf7" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.544521 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.562046 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.662302 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.680284 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573f7284-f946-46f2-9a4d-55ab2520a344-config-data\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.680760 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/573f7284-f946-46f2-9a4d-55ab2520a344-kolla-config\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.680792 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7lsn\" (UniqueName: \"kubernetes.io/projected/573f7284-f946-46f2-9a4d-55ab2520a344-kube-api-access-h7lsn\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.782004 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573f7284-f946-46f2-9a4d-55ab2520a344-config-data\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.782057 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/573f7284-f946-46f2-9a4d-55ab2520a344-kolla-config\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.782080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7lsn\" (UniqueName: \"kubernetes.io/projected/573f7284-f946-46f2-9a4d-55ab2520a344-kube-api-access-h7lsn\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.783407 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/573f7284-f946-46f2-9a4d-55ab2520a344-kolla-config\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.783692 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573f7284-f946-46f2-9a4d-55ab2520a344-config-data\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.823018 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7lsn\" (UniqueName: \"kubernetes.io/projected/573f7284-f946-46f2-9a4d-55ab2520a344-kube-api-access-h7lsn\") pod \"memcached-0\" (UID: \"573f7284-f946-46f2-9a4d-55ab2520a344\") " pod="openstack/memcached-0" Dec 05 12:27:46 crc kubenswrapper[4809]: I1205 12:27:46.861125 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.274094 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 12:27:47 crc kubenswrapper[4809]: W1205 12:27:47.278758 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5976c836_3783_481b_8dcb_8b5725a16e4c.slice/crio-77cc1bacd28f3e85929f7d0eb05530da664fb8726653ddddf46161a7e29a6b7f WatchSource:0}: Error finding container 77cc1bacd28f3e85929f7d0eb05530da664fb8726653ddddf46161a7e29a6b7f: Status 404 returned error can't find the container with id 77cc1bacd28f3e85929f7d0eb05530da664fb8726653ddddf46161a7e29a6b7f Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.367861 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 12:27:47 crc kubenswrapper[4809]: W1205 12:27:47.387531 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod573f7284_f946_46f2_9a4d_55ab2520a344.slice/crio-a64d59e2a4a0d06bf1c5617438785667aa097195bd8ad9cd8c5484e022c3a92d WatchSource:0}: Error finding container a64d59e2a4a0d06bf1c5617438785667aa097195bd8ad9cd8c5484e022c3a92d: Status 404 returned error can't find the container with id a64d59e2a4a0d06bf1c5617438785667aa097195bd8ad9cd8c5484e022c3a92d Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.404259 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87","Type":"ContainerStarted","Data":"4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1"} Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.407029 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5976c836-3783-481b-8dcb-8b5725a16e4c","Type":"ContainerStarted","Data":"77cc1bacd28f3e85929f7d0eb05530da664fb8726653ddddf46161a7e29a6b7f"} Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.408956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f15947f-ce8c-42b9-8221-aae51566ebf8","Type":"ContainerStarted","Data":"8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3"} Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.411072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"573f7284-f946-46f2-9a4d-55ab2520a344","Type":"ContainerStarted","Data":"a64d59e2a4a0d06bf1c5617438785667aa097195bd8ad9cd8c5484e022c3a92d"} Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.833291 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.834791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.837348 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-ktv5m" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.837867 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.838592 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.838955 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.860389 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.905905 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.906010 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl5vx\" (UniqueName: \"kubernetes.io/projected/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-kube-api-access-sl5vx\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.906101 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.906161 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.906200 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.906332 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.906426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:47 crc kubenswrapper[4809]: I1205 12:27:47.906505 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl5vx\" (UniqueName: \"kubernetes.io/projected/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-kube-api-access-sl5vx\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008275 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008301 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.008510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.009191 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.009327 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.009425 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.010744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.012300 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.012332 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d1b9c0097a4dc3ddd5e4ed68145e15caf55f0ede74b39557f96a634985187df2/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.014146 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.014425 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.033647 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl5vx\" (UniqueName: \"kubernetes.io/projected/295ed14b-b2f8-42eb-a9c8-1fc24e74654e-kube-api-access-sl5vx\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.043124 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d5b2407-458b-4fb2-87ec-150a5ba9450c\") pod \"openstack-cell1-galera-0\" (UID: \"295ed14b-b2f8-42eb-a9c8-1fc24e74654e\") " pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.153047 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.421383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"573f7284-f946-46f2-9a4d-55ab2520a344","Type":"ContainerStarted","Data":"d7066930c68b714cb59e73b8c5b435ea06b528a577aa8daaebdfa41952842cc2"} Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.422591 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.424642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5976c836-3783-481b-8dcb-8b5725a16e4c","Type":"ContainerStarted","Data":"d56f209934f7e394d7c08eb3d8f4deacf9c0945db7b2694e9606f2485b0d409e"} Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.445728 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.445705897 podStartE2EDuration="2.445705897s" podCreationTimestamp="2025-12-05 12:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:27:48.438848592 +0000 UTC m=+4763.829825160" watchObservedRunningTime="2025-12-05 12:27:48.445705897 +0000 UTC m=+4763.836682455" Dec 05 12:27:48 crc kubenswrapper[4809]: I1205 12:27:48.638722 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 12:27:48 crc kubenswrapper[4809]: W1205 12:27:48.648946 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod295ed14b_b2f8_42eb_a9c8_1fc24e74654e.slice/crio-be00421db7aae521eff81ee7d643f3e86f7c79490165aa3880ea785e286503ac WatchSource:0}: Error finding container be00421db7aae521eff81ee7d643f3e86f7c79490165aa3880ea785e286503ac: Status 404 returned error can't find the container with id be00421db7aae521eff81ee7d643f3e86f7c79490165aa3880ea785e286503ac Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.148786 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vbj8w"] Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.151062 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.167904 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbj8w"] Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.228306 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w6zc\" (UniqueName: \"kubernetes.io/projected/db163cca-30f5-47fa-8f04-4841858935d0-kube-api-access-5w6zc\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.228404 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-catalog-content\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.228450 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-utilities\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.330178 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w6zc\" (UniqueName: \"kubernetes.io/projected/db163cca-30f5-47fa-8f04-4841858935d0-kube-api-access-5w6zc\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.330473 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-catalog-content\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.330503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-utilities\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.330963 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-utilities\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.331051 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-catalog-content\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.353045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w6zc\" (UniqueName: \"kubernetes.io/projected/db163cca-30f5-47fa-8f04-4841858935d0-kube-api-access-5w6zc\") pod \"certified-operators-vbj8w\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.433754 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"295ed14b-b2f8-42eb-a9c8-1fc24e74654e","Type":"ContainerStarted","Data":"78249bcab265195116662a04927bfb79e86f5bcdb57b0aef48ce37e82217ceb2"} Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.434055 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"295ed14b-b2f8-42eb-a9c8-1fc24e74654e","Type":"ContainerStarted","Data":"be00421db7aae521eff81ee7d643f3e86f7c79490165aa3880ea785e286503ac"} Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.477117 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:49 crc kubenswrapper[4809]: I1205 12:27:49.989487 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbj8w"] Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.144414 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nltww"] Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.145894 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.169223 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nltww"] Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.319884 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-catalog-content\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.319998 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g5g7\" (UniqueName: \"kubernetes.io/projected/65ce5287-aef1-4b07-9c58-d9d38cce5720-kube-api-access-9g5g7\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.320028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-utilities\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.422105 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-catalog-content\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.422190 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g5g7\" (UniqueName: \"kubernetes.io/projected/65ce5287-aef1-4b07-9c58-d9d38cce5720-kube-api-access-9g5g7\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.422217 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-utilities\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.423884 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-utilities\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.424068 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-catalog-content\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.441828 4809 generic.go:334] "Generic (PLEG): container finished" podID="db163cca-30f5-47fa-8f04-4841858935d0" containerID="0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b" exitCode=0 Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.441901 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbj8w" event={"ID":"db163cca-30f5-47fa-8f04-4841858935d0","Type":"ContainerDied","Data":"0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b"} Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.441932 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbj8w" event={"ID":"db163cca-30f5-47fa-8f04-4841858935d0","Type":"ContainerStarted","Data":"c61d46f907b57d0d136c8f430b68a4138e210110bf3ad0743f4479c148b7af7d"} Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.458801 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g5g7\" (UniqueName: \"kubernetes.io/projected/65ce5287-aef1-4b07-9c58-d9d38cce5720-kube-api-access-9g5g7\") pod \"redhat-marketplace-nltww\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:50 crc kubenswrapper[4809]: I1205 12:27:50.471345 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:27:51 crc kubenswrapper[4809]: I1205 12:27:51.020489 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nltww"] Dec 05 12:27:51 crc kubenswrapper[4809]: W1205 12:27:51.023828 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65ce5287_aef1_4b07_9c58_d9d38cce5720.slice/crio-5e170a61d8b56d06686dd9bd4a3dc8f6877436bdbb06a8b7f9ac3ff29078f86f WatchSource:0}: Error finding container 5e170a61d8b56d06686dd9bd4a3dc8f6877436bdbb06a8b7f9ac3ff29078f86f: Status 404 returned error can't find the container with id 5e170a61d8b56d06686dd9bd4a3dc8f6877436bdbb06a8b7f9ac3ff29078f86f Dec 05 12:27:51 crc kubenswrapper[4809]: I1205 12:27:51.453068 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbj8w" event={"ID":"db163cca-30f5-47fa-8f04-4841858935d0","Type":"ContainerStarted","Data":"c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed"} Dec 05 12:27:51 crc kubenswrapper[4809]: I1205 12:27:51.455104 4809 generic.go:334] "Generic (PLEG): container finished" podID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerID="ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30" exitCode=0 Dec 05 12:27:51 crc kubenswrapper[4809]: I1205 12:27:51.455183 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nltww" event={"ID":"65ce5287-aef1-4b07-9c58-d9d38cce5720","Type":"ContainerDied","Data":"ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30"} Dec 05 12:27:51 crc kubenswrapper[4809]: I1205 12:27:51.455415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nltww" event={"ID":"65ce5287-aef1-4b07-9c58-d9d38cce5720","Type":"ContainerStarted","Data":"5e170a61d8b56d06686dd9bd4a3dc8f6877436bdbb06a8b7f9ac3ff29078f86f"} Dec 05 12:27:51 crc kubenswrapper[4809]: I1205 12:27:51.457505 4809 generic.go:334] "Generic (PLEG): container finished" podID="5976c836-3783-481b-8dcb-8b5725a16e4c" containerID="d56f209934f7e394d7c08eb3d8f4deacf9c0945db7b2694e9606f2485b0d409e" exitCode=0 Dec 05 12:27:51 crc kubenswrapper[4809]: I1205 12:27:51.457548 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5976c836-3783-481b-8dcb-8b5725a16e4c","Type":"ContainerDied","Data":"d56f209934f7e394d7c08eb3d8f4deacf9c0945db7b2694e9606f2485b0d409e"} Dec 05 12:27:52 crc kubenswrapper[4809]: I1205 12:27:52.467575 4809 generic.go:334] "Generic (PLEG): container finished" podID="db163cca-30f5-47fa-8f04-4841858935d0" containerID="c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed" exitCode=0 Dec 05 12:27:52 crc kubenswrapper[4809]: I1205 12:27:52.467650 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbj8w" event={"ID":"db163cca-30f5-47fa-8f04-4841858935d0","Type":"ContainerDied","Data":"c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed"} Dec 05 12:27:52 crc kubenswrapper[4809]: I1205 12:27:52.471413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nltww" event={"ID":"65ce5287-aef1-4b07-9c58-d9d38cce5720","Type":"ContainerStarted","Data":"624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6"} Dec 05 12:27:52 crc kubenswrapper[4809]: I1205 12:27:52.473449 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5976c836-3783-481b-8dcb-8b5725a16e4c","Type":"ContainerStarted","Data":"7bfcea81892c757414a5798f94886ccfebf04ce1911cf2cc5e3b501cb9347d70"} Dec 05 12:27:52 crc kubenswrapper[4809]: I1205 12:27:52.509055 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.509034611 podStartE2EDuration="8.509034611s" podCreationTimestamp="2025-12-05 12:27:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:27:52.507333945 +0000 UTC m=+4767.898310513" watchObservedRunningTime="2025-12-05 12:27:52.509034611 +0000 UTC m=+4767.900011169" Dec 05 12:27:53 crc kubenswrapper[4809]: I1205 12:27:53.482488 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbj8w" event={"ID":"db163cca-30f5-47fa-8f04-4841858935d0","Type":"ContainerStarted","Data":"f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc"} Dec 05 12:27:53 crc kubenswrapper[4809]: I1205 12:27:53.484439 4809 generic.go:334] "Generic (PLEG): container finished" podID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerID="624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6" exitCode=0 Dec 05 12:27:53 crc kubenswrapper[4809]: I1205 12:27:53.484530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nltww" event={"ID":"65ce5287-aef1-4b07-9c58-d9d38cce5720","Type":"ContainerDied","Data":"624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6"} Dec 05 12:27:53 crc kubenswrapper[4809]: I1205 12:27:53.485832 4809 generic.go:334] "Generic (PLEG): container finished" podID="295ed14b-b2f8-42eb-a9c8-1fc24e74654e" containerID="78249bcab265195116662a04927bfb79e86f5bcdb57b0aef48ce37e82217ceb2" exitCode=0 Dec 05 12:27:53 crc kubenswrapper[4809]: I1205 12:27:53.485870 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"295ed14b-b2f8-42eb-a9c8-1fc24e74654e","Type":"ContainerDied","Data":"78249bcab265195116662a04927bfb79e86f5bcdb57b0aef48ce37e82217ceb2"} Dec 05 12:27:53 crc kubenswrapper[4809]: I1205 12:27:53.511195 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vbj8w" podStartSLOduration=2.127523867 podStartE2EDuration="4.511172338s" podCreationTimestamp="2025-12-05 12:27:49 +0000 UTC" firstStartedPulling="2025-12-05 12:27:50.443854896 +0000 UTC m=+4765.834831454" lastFinishedPulling="2025-12-05 12:27:52.827503367 +0000 UTC m=+4768.218479925" observedRunningTime="2025-12-05 12:27:53.506299537 +0000 UTC m=+4768.897276145" watchObservedRunningTime="2025-12-05 12:27:53.511172338 +0000 UTC m=+4768.902148906" Dec 05 12:27:53 crc kubenswrapper[4809]: I1205 12:27:53.898937 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:54 crc kubenswrapper[4809]: I1205 12:27:54.191784 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:27:54 crc kubenswrapper[4809]: I1205 12:27:54.258607 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-ptwr6"] Dec 05 12:27:54 crc kubenswrapper[4809]: I1205 12:27:54.492847 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"295ed14b-b2f8-42eb-a9c8-1fc24e74654e","Type":"ContainerStarted","Data":"9d305734e9f7285325b2947629b5552f38ca991524e0ee66a9f170a06bceba1b"} Dec 05 12:27:54 crc kubenswrapper[4809]: I1205 12:27:54.495213 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nltww" event={"ID":"65ce5287-aef1-4b07-9c58-d9d38cce5720","Type":"ContainerStarted","Data":"a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541"} Dec 05 12:27:54 crc kubenswrapper[4809]: I1205 12:27:54.495721 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" podUID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerName="dnsmasq-dns" containerID="cri-o://738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1" gracePeriod=10 Dec 05 12:27:54 crc kubenswrapper[4809]: I1205 12:27:54.544695 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.544619479 podStartE2EDuration="8.544619479s" podCreationTimestamp="2025-12-05 12:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:27:54.527105367 +0000 UTC m=+4769.918081925" watchObservedRunningTime="2025-12-05 12:27:54.544619479 +0000 UTC m=+4769.935596047" Dec 05 12:27:54 crc kubenswrapper[4809]: I1205 12:27:54.549726 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nltww" podStartSLOduration=2.137927307 podStartE2EDuration="4.549705896s" podCreationTimestamp="2025-12-05 12:27:50 +0000 UTC" firstStartedPulling="2025-12-05 12:27:51.456743673 +0000 UTC m=+4766.847720241" lastFinishedPulling="2025-12-05 12:27:53.868522262 +0000 UTC m=+4769.259498830" observedRunningTime="2025-12-05 12:27:54.54355759 +0000 UTC m=+4769.934534148" watchObservedRunningTime="2025-12-05 12:27:54.549705896 +0000 UTC m=+4769.940682464" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.437564 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.507652 4809 generic.go:334] "Generic (PLEG): container finished" podID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerID="738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1" exitCode=0 Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.508415 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" event={"ID":"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e","Type":"ContainerDied","Data":"738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1"} Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.508462 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" event={"ID":"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e","Type":"ContainerDied","Data":"7561176eb2179661388f6c826d05144ae550c9289ef709123648a7212dcc40a3"} Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.508484 4809 scope.go:117] "RemoveContainer" containerID="738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.508487 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-ptwr6" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.530227 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-config\") pod \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.530328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-dns-svc\") pod \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.530429 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4vvc\" (UniqueName: \"kubernetes.io/projected/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-kube-api-access-q4vvc\") pod \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\" (UID: \"51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e\") " Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.541405 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-kube-api-access-q4vvc" (OuterVolumeSpecName: "kube-api-access-q4vvc") pod "51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" (UID: "51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e"). InnerVolumeSpecName "kube-api-access-q4vvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.562424 4809 scope.go:117] "RemoveContainer" containerID="d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.580233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" (UID: "51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.592922 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-config" (OuterVolumeSpecName: "config") pod "51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" (UID: "51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.604042 4809 scope.go:117] "RemoveContainer" containerID="738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1" Dec 05 12:27:55 crc kubenswrapper[4809]: E1205 12:27:55.604552 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1\": container with ID starting with 738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1 not found: ID does not exist" containerID="738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.604588 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1"} err="failed to get container status \"738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1\": rpc error: code = NotFound desc = could not find container \"738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1\": container with ID starting with 738ef84956a5276b45cacc1aed6bbef0d0fa40bdc167b596e3ebaa97ad71c8d1 not found: ID does not exist" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.604612 4809 scope.go:117] "RemoveContainer" containerID="d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6" Dec 05 12:27:55 crc kubenswrapper[4809]: E1205 12:27:55.605039 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6\": container with ID starting with d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6 not found: ID does not exist" containerID="d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.605076 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6"} err="failed to get container status \"d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6\": rpc error: code = NotFound desc = could not find container \"d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6\": container with ID starting with d4a5ee83b6d66df9c36b405aa93bc5d4a6dce8114a8aaf1804d0874a1652c8a6 not found: ID does not exist" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.632040 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.632074 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.632084 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4vvc\" (UniqueName: \"kubernetes.io/projected/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e-kube-api-access-q4vvc\") on node \"crc\" DevicePath \"\"" Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.842028 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-ptwr6"] Dec 05 12:27:55 crc kubenswrapper[4809]: I1205 12:27:55.847733 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-ptwr6"] Dec 05 12:27:56 crc kubenswrapper[4809]: I1205 12:27:56.663479 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 05 12:27:56 crc kubenswrapper[4809]: I1205 12:27:56.663530 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 05 12:27:56 crc kubenswrapper[4809]: I1205 12:27:56.862513 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 05 12:27:56 crc kubenswrapper[4809]: I1205 12:27:56.892549 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" path="/var/lib/kubelet/pods/51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e/volumes" Dec 05 12:27:58 crc kubenswrapper[4809]: I1205 12:27:58.153578 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:58 crc kubenswrapper[4809]: I1205 12:27:58.153657 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 05 12:27:58 crc kubenswrapper[4809]: I1205 12:27:58.833692 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 05 12:27:59 crc kubenswrapper[4809]: I1205 12:27:59.046149 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 05 12:27:59 crc kubenswrapper[4809]: I1205 12:27:59.477698 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:59 crc kubenswrapper[4809]: I1205 12:27:59.477752 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:59 crc kubenswrapper[4809]: I1205 12:27:59.522089 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:59 crc kubenswrapper[4809]: I1205 12:27:59.580811 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:27:59 crc kubenswrapper[4809]: I1205 12:27:59.938010 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbj8w"] Dec 05 12:28:00 crc kubenswrapper[4809]: I1205 12:28:00.471881 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:28:00 crc kubenswrapper[4809]: I1205 12:28:00.472226 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:28:00 crc kubenswrapper[4809]: I1205 12:28:00.516474 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:28:00 crc kubenswrapper[4809]: I1205 12:28:00.587240 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:28:00 crc kubenswrapper[4809]: I1205 12:28:00.651606 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 05 12:28:00 crc kubenswrapper[4809]: I1205 12:28:00.730446 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 05 12:28:01 crc kubenswrapper[4809]: I1205 12:28:01.553176 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vbj8w" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="registry-server" containerID="cri-o://f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc" gracePeriod=2 Dec 05 12:28:01 crc kubenswrapper[4809]: I1205 12:28:01.964546 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.040664 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w6zc\" (UniqueName: \"kubernetes.io/projected/db163cca-30f5-47fa-8f04-4841858935d0-kube-api-access-5w6zc\") pod \"db163cca-30f5-47fa-8f04-4841858935d0\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.040800 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-catalog-content\") pod \"db163cca-30f5-47fa-8f04-4841858935d0\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.040913 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-utilities\") pod \"db163cca-30f5-47fa-8f04-4841858935d0\" (UID: \"db163cca-30f5-47fa-8f04-4841858935d0\") " Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.042246 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-utilities" (OuterVolumeSpecName: "utilities") pod "db163cca-30f5-47fa-8f04-4841858935d0" (UID: "db163cca-30f5-47fa-8f04-4841858935d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.048006 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db163cca-30f5-47fa-8f04-4841858935d0-kube-api-access-5w6zc" (OuterVolumeSpecName: "kube-api-access-5w6zc") pod "db163cca-30f5-47fa-8f04-4841858935d0" (UID: "db163cca-30f5-47fa-8f04-4841858935d0"). InnerVolumeSpecName "kube-api-access-5w6zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.142403 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.142451 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w6zc\" (UniqueName: \"kubernetes.io/projected/db163cca-30f5-47fa-8f04-4841858935d0-kube-api-access-5w6zc\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.195313 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db163cca-30f5-47fa-8f04-4841858935d0" (UID: "db163cca-30f5-47fa-8f04-4841858935d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.244128 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db163cca-30f5-47fa-8f04-4841858935d0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.537228 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nltww"] Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.569873 4809 generic.go:334] "Generic (PLEG): container finished" podID="db163cca-30f5-47fa-8f04-4841858935d0" containerID="f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc" exitCode=0 Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.570432 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nltww" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="registry-server" containerID="cri-o://a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541" gracePeriod=2 Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.570733 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbj8w" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.570866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbj8w" event={"ID":"db163cca-30f5-47fa-8f04-4841858935d0","Type":"ContainerDied","Data":"f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc"} Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.570972 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbj8w" event={"ID":"db163cca-30f5-47fa-8f04-4841858935d0","Type":"ContainerDied","Data":"c61d46f907b57d0d136c8f430b68a4138e210110bf3ad0743f4479c148b7af7d"} Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.571015 4809 scope.go:117] "RemoveContainer" containerID="f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.596383 4809 scope.go:117] "RemoveContainer" containerID="c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.616589 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbj8w"] Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.622090 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vbj8w"] Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.644462 4809 scope.go:117] "RemoveContainer" containerID="0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.678605 4809 scope.go:117] "RemoveContainer" containerID="f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc" Dec 05 12:28:02 crc kubenswrapper[4809]: E1205 12:28:02.679965 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc\": container with ID starting with f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc not found: ID does not exist" containerID="f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.680107 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc"} err="failed to get container status \"f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc\": rpc error: code = NotFound desc = could not find container \"f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc\": container with ID starting with f6bf3163d00de737e5d1d731ca8e00e9942393d2130b3dd3c4377c5ff2239ccc not found: ID does not exist" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.680222 4809 scope.go:117] "RemoveContainer" containerID="c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed" Dec 05 12:28:02 crc kubenswrapper[4809]: E1205 12:28:02.680791 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed\": container with ID starting with c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed not found: ID does not exist" containerID="c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.680919 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed"} err="failed to get container status \"c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed\": rpc error: code = NotFound desc = could not find container \"c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed\": container with ID starting with c34dfe6e8dbf14b8837cb0fcabcdd157edcee8350229e264949071548fb19eed not found: ID does not exist" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.681009 4809 scope.go:117] "RemoveContainer" containerID="0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b" Dec 05 12:28:02 crc kubenswrapper[4809]: E1205 12:28:02.681440 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b\": container with ID starting with 0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b not found: ID does not exist" containerID="0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.681468 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b"} err="failed to get container status \"0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b\": rpc error: code = NotFound desc = could not find container \"0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b\": container with ID starting with 0098fab65e9e37dba85cb1b02c1cc5504769b06f686fa63b526680ce839d8a3b not found: ID does not exist" Dec 05 12:28:02 crc kubenswrapper[4809]: I1205 12:28:02.881532 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db163cca-30f5-47fa-8f04-4841858935d0" path="/var/lib/kubelet/pods/db163cca-30f5-47fa-8f04-4841858935d0/volumes" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.480139 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.562008 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-utilities\") pod \"65ce5287-aef1-4b07-9c58-d9d38cce5720\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.562058 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-catalog-content\") pod \"65ce5287-aef1-4b07-9c58-d9d38cce5720\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.562116 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g5g7\" (UniqueName: \"kubernetes.io/projected/65ce5287-aef1-4b07-9c58-d9d38cce5720-kube-api-access-9g5g7\") pod \"65ce5287-aef1-4b07-9c58-d9d38cce5720\" (UID: \"65ce5287-aef1-4b07-9c58-d9d38cce5720\") " Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.563395 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-utilities" (OuterVolumeSpecName: "utilities") pod "65ce5287-aef1-4b07-9c58-d9d38cce5720" (UID: "65ce5287-aef1-4b07-9c58-d9d38cce5720"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.565653 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ce5287-aef1-4b07-9c58-d9d38cce5720-kube-api-access-9g5g7" (OuterVolumeSpecName: "kube-api-access-9g5g7") pod "65ce5287-aef1-4b07-9c58-d9d38cce5720" (UID: "65ce5287-aef1-4b07-9c58-d9d38cce5720"). InnerVolumeSpecName "kube-api-access-9g5g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.580586 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65ce5287-aef1-4b07-9c58-d9d38cce5720" (UID: "65ce5287-aef1-4b07-9c58-d9d38cce5720"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.581161 4809 generic.go:334] "Generic (PLEG): container finished" podID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerID="a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541" exitCode=0 Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.581238 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nltww" event={"ID":"65ce5287-aef1-4b07-9c58-d9d38cce5720","Type":"ContainerDied","Data":"a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541"} Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.581246 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nltww" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.581284 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nltww" event={"ID":"65ce5287-aef1-4b07-9c58-d9d38cce5720","Type":"ContainerDied","Data":"5e170a61d8b56d06686dd9bd4a3dc8f6877436bdbb06a8b7f9ac3ff29078f86f"} Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.581309 4809 scope.go:117] "RemoveContainer" containerID="a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.610869 4809 scope.go:117] "RemoveContainer" containerID="624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.625603 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nltww"] Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.632065 4809 scope.go:117] "RemoveContainer" containerID="ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.633872 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nltww"] Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.646841 4809 scope.go:117] "RemoveContainer" containerID="a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541" Dec 05 12:28:03 crc kubenswrapper[4809]: E1205 12:28:03.647193 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541\": container with ID starting with a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541 not found: ID does not exist" containerID="a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.647226 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541"} err="failed to get container status \"a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541\": rpc error: code = NotFound desc = could not find container \"a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541\": container with ID starting with a1ff7c175ff8629ff5cbdff8a35ee16d322c4f0596a2bccfbabfa00a8d58b541 not found: ID does not exist" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.647247 4809 scope.go:117] "RemoveContainer" containerID="624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6" Dec 05 12:28:03 crc kubenswrapper[4809]: E1205 12:28:03.647714 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6\": container with ID starting with 624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6 not found: ID does not exist" containerID="624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.647741 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6"} err="failed to get container status \"624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6\": rpc error: code = NotFound desc = could not find container \"624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6\": container with ID starting with 624e8ea0cf4da60f6af84997917598418da0f54c70b3e8843140dcbec81040a6 not found: ID does not exist" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.647756 4809 scope.go:117] "RemoveContainer" containerID="ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30" Dec 05 12:28:03 crc kubenswrapper[4809]: E1205 12:28:03.647951 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30\": container with ID starting with ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30 not found: ID does not exist" containerID="ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.647973 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30"} err="failed to get container status \"ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30\": rpc error: code = NotFound desc = could not find container \"ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30\": container with ID starting with ed2f995b4f4ab17582add6f86d6d8f8e7a35fd2667ba353096815295f5526b30 not found: ID does not exist" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.663307 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g5g7\" (UniqueName: \"kubernetes.io/projected/65ce5287-aef1-4b07-9c58-d9d38cce5720-kube-api-access-9g5g7\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.663333 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:03 crc kubenswrapper[4809]: I1205 12:28:03.663341 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65ce5287-aef1-4b07-9c58-d9d38cce5720-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:04 crc kubenswrapper[4809]: I1205 12:28:04.882974 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" path="/var/lib/kubelet/pods/65ce5287-aef1-4b07-9c58-d9d38cce5720/volumes" Dec 05 12:28:14 crc kubenswrapper[4809]: I1205 12:28:14.047028 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:28:14 crc kubenswrapper[4809]: I1205 12:28:14.047683 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:28:18 crc kubenswrapper[4809]: I1205 12:28:18.716178 4809 generic.go:334] "Generic (PLEG): container finished" podID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerID="8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3" exitCode=0 Dec 05 12:28:18 crc kubenswrapper[4809]: I1205 12:28:18.716277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f15947f-ce8c-42b9-8221-aae51566ebf8","Type":"ContainerDied","Data":"8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3"} Dec 05 12:28:19 crc kubenswrapper[4809]: I1205 12:28:19.727711 4809 generic.go:334] "Generic (PLEG): container finished" podID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerID="4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1" exitCode=0 Dec 05 12:28:19 crc kubenswrapper[4809]: I1205 12:28:19.727780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87","Type":"ContainerDied","Data":"4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1"} Dec 05 12:28:19 crc kubenswrapper[4809]: I1205 12:28:19.730413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f15947f-ce8c-42b9-8221-aae51566ebf8","Type":"ContainerStarted","Data":"a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34"} Dec 05 12:28:19 crc kubenswrapper[4809]: I1205 12:28:19.730674 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 12:28:19 crc kubenswrapper[4809]: I1205 12:28:19.798861 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.798842361 podStartE2EDuration="36.798842361s" podCreationTimestamp="2025-12-05 12:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:28:19.796389895 +0000 UTC m=+4795.187366473" watchObservedRunningTime="2025-12-05 12:28:19.798842361 +0000 UTC m=+4795.189818919" Dec 05 12:28:20 crc kubenswrapper[4809]: I1205 12:28:20.741423 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87","Type":"ContainerStarted","Data":"923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3"} Dec 05 12:28:20 crc kubenswrapper[4809]: I1205 12:28:20.742440 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:20 crc kubenswrapper[4809]: I1205 12:28:20.771930 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.771905433 podStartE2EDuration="37.771905433s" podCreationTimestamp="2025-12-05 12:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:28:20.767978407 +0000 UTC m=+4796.158954995" watchObservedRunningTime="2025-12-05 12:28:20.771905433 +0000 UTC m=+4796.162882011" Dec 05 12:28:35 crc kubenswrapper[4809]: I1205 12:28:35.255715 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 12:28:35 crc kubenswrapper[4809]: I1205 12:28:35.363853 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.394895 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-4zfn2"] Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.395872 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerName="dnsmasq-dns" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.395892 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerName="dnsmasq-dns" Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.395913 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="registry-server" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.395920 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="registry-server" Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.395951 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="registry-server" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.395960 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="registry-server" Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.395976 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="extract-utilities" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.395984 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="extract-utilities" Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.395994 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="extract-utilities" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.396001 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="extract-utilities" Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.396022 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="extract-content" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.396029 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="extract-content" Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.396044 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="extract-content" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.396052 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="extract-content" Dec 05 12:28:40 crc kubenswrapper[4809]: E1205 12:28:40.396066 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerName="init" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.396073 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerName="init" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.396246 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ce5287-aef1-4b07-9c58-d9d38cce5720" containerName="registry-server" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.396282 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="51dfdfbc-61ff-44c7-bbc2-ec8a3970c34e" containerName="dnsmasq-dns" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.396298 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="db163cca-30f5-47fa-8f04-4841858935d0" containerName="registry-server" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.397355 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.418110 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-4zfn2"] Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.443834 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snzld\" (UniqueName: \"kubernetes.io/projected/bd51984c-a4ab-4041-8586-94c4fe70ab60-kube-api-access-snzld\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.443914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-config\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.443955 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.545240 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snzld\" (UniqueName: \"kubernetes.io/projected/bd51984c-a4ab-4041-8586-94c4fe70ab60-kube-api-access-snzld\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.545367 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-config\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.545441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.546525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-config\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.546934 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.565063 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snzld\" (UniqueName: \"kubernetes.io/projected/bd51984c-a4ab-4041-8586-94c4fe70ab60-kube-api-access-snzld\") pod \"dnsmasq-dns-5b7946d7b9-4zfn2\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.718369 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:40 crc kubenswrapper[4809]: I1205 12:28:40.957715 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-4zfn2"] Dec 05 12:28:41 crc kubenswrapper[4809]: I1205 12:28:41.110387 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:28:41 crc kubenswrapper[4809]: I1205 12:28:41.903852 4809 generic.go:334] "Generic (PLEG): container finished" podID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerID="b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec" exitCode=0 Dec 05 12:28:41 crc kubenswrapper[4809]: I1205 12:28:41.903956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" event={"ID":"bd51984c-a4ab-4041-8586-94c4fe70ab60","Type":"ContainerDied","Data":"b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec"} Dec 05 12:28:41 crc kubenswrapper[4809]: I1205 12:28:41.904165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" event={"ID":"bd51984c-a4ab-4041-8586-94c4fe70ab60","Type":"ContainerStarted","Data":"c20b9e7ec248a5dc02dd44cd0b2b5da290b877fb6ced23a50fa4376b3f3d778b"} Dec 05 12:28:41 crc kubenswrapper[4809]: I1205 12:28:41.920350 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:28:42 crc kubenswrapper[4809]: I1205 12:28:42.920391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" event={"ID":"bd51984c-a4ab-4041-8586-94c4fe70ab60","Type":"ContainerStarted","Data":"3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa"} Dec 05 12:28:42 crc kubenswrapper[4809]: I1205 12:28:42.920837 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:42 crc kubenswrapper[4809]: I1205 12:28:42.946814 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" podStartSLOduration=2.9467818489999997 podStartE2EDuration="2.946781849s" podCreationTimestamp="2025-12-05 12:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:28:42.943166841 +0000 UTC m=+4818.334143399" watchObservedRunningTime="2025-12-05 12:28:42.946781849 +0000 UTC m=+4818.337758447" Dec 05 12:28:43 crc kubenswrapper[4809]: I1205 12:28:43.007160 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerName="rabbitmq" containerID="cri-o://a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34" gracePeriod=604799 Dec 05 12:28:43 crc kubenswrapper[4809]: I1205 12:28:43.873735 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerName="rabbitmq" containerID="cri-o://923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3" gracePeriod=604799 Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.046748 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.046810 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.046856 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.047473 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90fcfc02adb974dcd81059abed9da13265a950cbccc4646c761fd364738d11bd"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.047532 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://90fcfc02adb974dcd81059abed9da13265a950cbccc4646c761fd364738d11bd" gracePeriod=600 Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.937898 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="90fcfc02adb974dcd81059abed9da13265a950cbccc4646c761fd364738d11bd" exitCode=0 Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.938023 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"90fcfc02adb974dcd81059abed9da13265a950cbccc4646c761fd364738d11bd"} Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.938367 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02"} Dec 05 12:28:44 crc kubenswrapper[4809]: I1205 12:28:44.938389 4809 scope.go:117] "RemoveContainer" containerID="35725e9be725c53c9013a58e97f01d0dffc341d89593f04a99771a0f527922d8" Dec 05 12:28:45 crc kubenswrapper[4809]: I1205 12:28:45.254140 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.238:5672: connect: connection refused" Dec 05 12:28:45 crc kubenswrapper[4809]: I1205 12:28:45.360045 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.239:5672: connect: connection refused" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.539601 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.688071 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-erlang-cookie\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.688462 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-plugins\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.688491 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2klf\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-kube-api-access-v2klf\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.688527 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f15947f-ce8c-42b9-8221-aae51566ebf8-pod-info\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.688619 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f15947f-ce8c-42b9-8221-aae51566ebf8-erlang-cookie-secret\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.688697 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-confd\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.688885 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.689075 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.689550 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.689598 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-server-conf\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.689618 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-plugins-conf\") pod \"5f15947f-ce8c-42b9-8221-aae51566ebf8\" (UID: \"5f15947f-ce8c-42b9-8221-aae51566ebf8\") " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.690389 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.690450 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.690478 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.695169 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f15947f-ce8c-42b9-8221-aae51566ebf8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.695301 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-kube-api-access-v2klf" (OuterVolumeSpecName: "kube-api-access-v2klf") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "kube-api-access-v2klf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.702247 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5f15947f-ce8c-42b9-8221-aae51566ebf8-pod-info" (OuterVolumeSpecName: "pod-info") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.712351 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d" (OuterVolumeSpecName: "persistence") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.718619 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-server-conf" (OuterVolumeSpecName: "server-conf") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.768940 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5f15947f-ce8c-42b9-8221-aae51566ebf8" (UID: "5f15947f-ce8c-42b9-8221-aae51566ebf8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.791684 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") on node \"crc\" " Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.791718 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.791728 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f15947f-ce8c-42b9-8221-aae51566ebf8-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.791738 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2klf\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-kube-api-access-v2klf\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.791749 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f15947f-ce8c-42b9-8221-aae51566ebf8-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.791757 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f15947f-ce8c-42b9-8221-aae51566ebf8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.791766 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f15947f-ce8c-42b9-8221-aae51566ebf8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.807414 4809 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.807545 4809 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d") on node "crc" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.893270 4809 reconciler_common.go:293] "Volume detached for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.992543 4809 generic.go:334] "Generic (PLEG): container finished" podID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerID="a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34" exitCode=0 Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.992589 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f15947f-ce8c-42b9-8221-aae51566ebf8","Type":"ContainerDied","Data":"a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34"} Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.992620 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f15947f-ce8c-42b9-8221-aae51566ebf8","Type":"ContainerDied","Data":"79b522386ce8bfed4c2549b97775bc1ee3f80b7a54d1d97d2a29f6a0d7e03e21"} Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.992657 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 12:28:49 crc kubenswrapper[4809]: I1205 12:28:49.992675 4809 scope.go:117] "RemoveContainer" containerID="a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.017611 4809 scope.go:117] "RemoveContainer" containerID="8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.060830 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.065108 4809 scope.go:117] "RemoveContainer" containerID="a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34" Dec 05 12:28:50 crc kubenswrapper[4809]: E1205 12:28:50.065667 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34\": container with ID starting with a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34 not found: ID does not exist" containerID="a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.065780 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34"} err="failed to get container status \"a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34\": rpc error: code = NotFound desc = could not find container \"a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34\": container with ID starting with a6b95a33b13bb1a47b04247d290ab5aa435a255cc028a9b5b92fad114309aa34 not found: ID does not exist" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.065861 4809 scope.go:117] "RemoveContainer" containerID="8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.066140 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:28:50 crc kubenswrapper[4809]: E1205 12:28:50.066367 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3\": container with ID starting with 8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3 not found: ID does not exist" containerID="8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.066398 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3"} err="failed to get container status \"8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3\": rpc error: code = NotFound desc = could not find container \"8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3\": container with ID starting with 8e36f9f3b4fa93893567677d1504e41b1b972f19431577a0ec7f8a5628bd93b3 not found: ID does not exist" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.076172 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:28:50 crc kubenswrapper[4809]: E1205 12:28:50.078975 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerName="rabbitmq" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.079103 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerName="rabbitmq" Dec 05 12:28:50 crc kubenswrapper[4809]: E1205 12:28:50.079205 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerName="setup-container" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.079296 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerName="setup-container" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.079587 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" containerName="rabbitmq" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.080765 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.082769 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.085793 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.087347 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.087498 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7pfpn" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.087615 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.087750 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.202050 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5b994bc1-9a84-41c6-97df-6ee239d92968-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.202461 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.202598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm9p6\" (UniqueName: \"kubernetes.io/projected/5b994bc1-9a84-41c6-97df-6ee239d92968-kube-api-access-pm9p6\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.202667 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.202885 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.202961 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5b994bc1-9a84-41c6-97df-6ee239d92968-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.203002 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5b994bc1-9a84-41c6-97df-6ee239d92968-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.203088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.203197 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5b994bc1-9a84-41c6-97df-6ee239d92968-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm9p6\" (UniqueName: \"kubernetes.io/projected/5b994bc1-9a84-41c6-97df-6ee239d92968-kube-api-access-pm9p6\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304598 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304627 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304681 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5b994bc1-9a84-41c6-97df-6ee239d92968-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304697 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5b994bc1-9a84-41c6-97df-6ee239d92968-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304726 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304768 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5b994bc1-9a84-41c6-97df-6ee239d92968-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5b994bc1-9a84-41c6-97df-6ee239d92968-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.304888 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.305423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.306073 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.306135 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5b994bc1-9a84-41c6-97df-6ee239d92968-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.306368 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5b994bc1-9a84-41c6-97df-6ee239d92968-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.307336 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.307368 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e22c51975a412c7ed096050c61fe20a6ae4ba86f0578c4586a4dfb3a62e17ae5/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.308764 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5b994bc1-9a84-41c6-97df-6ee239d92968-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.308770 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5b994bc1-9a84-41c6-97df-6ee239d92968-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.308857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5b994bc1-9a84-41c6-97df-6ee239d92968-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.322402 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm9p6\" (UniqueName: \"kubernetes.io/projected/5b994bc1-9a84-41c6-97df-6ee239d92968-kube-api-access-pm9p6\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.337546 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c6a3e5-dc03-483f-998d-db7020e89e7d\") pod \"rabbitmq-server-0\" (UID: \"5b994bc1-9a84-41c6-97df-6ee239d92968\") " pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.407549 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.721204 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.781311 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-9gdsv"] Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.781530 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" podUID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerName="dnsmasq-dns" containerID="cri-o://adc6b1eedf990527e5509a77d0678ce0582266094a8fb061799998318c6d30f0" gracePeriod=10 Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.889728 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f15947f-ce8c-42b9-8221-aae51566ebf8" path="/var/lib/kubelet/pods/5f15947f-ce8c-42b9-8221-aae51566ebf8/volumes" Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.897972 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 12:28:50 crc kubenswrapper[4809]: I1205 12:28:50.942338 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.017698 4809 generic.go:334] "Generic (PLEG): container finished" podID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerID="923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3" exitCode=0 Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.017767 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87","Type":"ContainerDied","Data":"923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3"} Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.017794 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87","Type":"ContainerDied","Data":"8356bcb35137dfec5fed33e64c3b88654e34459708782f7e4ebaf3469200a585"} Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.017811 4809 scope.go:117] "RemoveContainer" containerID="923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.017884 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.029441 4809 generic.go:334] "Generic (PLEG): container finished" podID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerID="adc6b1eedf990527e5509a77d0678ce0582266094a8fb061799998318c6d30f0" exitCode=0 Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.029541 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" event={"ID":"fde5555e-3d40-4b92-ab47-f9b2043ddb1b","Type":"ContainerDied","Data":"adc6b1eedf990527e5509a77d0678ce0582266094a8fb061799998318c6d30f0"} Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.038911 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5b994bc1-9a84-41c6-97df-6ee239d92968","Type":"ContainerStarted","Data":"2d608fbe7c5a489752b4bfd9f39eaed89bbe1907ef93354dc79b731329f16b4b"} Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.051783 4809 scope.go:117] "RemoveContainer" containerID="4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.070710 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-server-conf\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.071147 4809 scope.go:117] "RemoveContainer" containerID="923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3" Dec 05 12:28:51 crc kubenswrapper[4809]: E1205 12:28:51.076738 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3\": container with ID starting with 923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3 not found: ID does not exist" containerID="923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.076774 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3"} err="failed to get container status \"923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3\": rpc error: code = NotFound desc = could not find container \"923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3\": container with ID starting with 923a8e82d10ea87699f42ddbb732b314936f3653074e97e4e3310b45b1c64fc3 not found: ID does not exist" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.076804 4809 scope.go:117] "RemoveContainer" containerID="4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.076797 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dgk6\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-kube-api-access-2dgk6\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.076874 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-pod-info\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.076940 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-erlang-cookie\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.077005 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-plugins-conf\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.077037 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-confd\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.077199 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.077238 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-plugins\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.077307 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-erlang-cookie-secret\") pod \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\" (UID: \"ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87\") " Dec 05 12:28:51 crc kubenswrapper[4809]: E1205 12:28:51.079771 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1\": container with ID starting with 4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1 not found: ID does not exist" containerID="4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.079806 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1"} err="failed to get container status \"4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1\": rpc error: code = NotFound desc = could not find container \"4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1\": container with ID starting with 4422855870dbc897decdc23c61136641855098b790e129b9979208cb581325b1 not found: ID does not exist" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.080497 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.081079 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.082289 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.083892 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-kube-api-access-2dgk6" (OuterVolumeSpecName: "kube-api-access-2dgk6") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "kube-api-access-2dgk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.087346 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-pod-info" (OuterVolumeSpecName: "pod-info") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.087608 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.091809 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-server-conf" (OuterVolumeSpecName: "server-conf") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.094603 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721" (OuterVolumeSpecName: "persistence") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179880 4809 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179915 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dgk6\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-kube-api-access-2dgk6\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179927 4809 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179936 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179944 4809 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179975 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") on node \"crc\" " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179987 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.179996 4809 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.184108 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" (UID: "ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.197779 4809 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.197963 4809 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721") on node "crc" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.203678 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.281203 4809 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.281242 4809 reconciler_common.go:293] "Volume detached for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.351593 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.360875 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.381948 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-dns-svc\") pod \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.382002 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-config\") pod \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.382083 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzvsw\" (UniqueName: \"kubernetes.io/projected/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-kube-api-access-xzvsw\") pod \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\" (UID: \"fde5555e-3d40-4b92-ab47-f9b2043ddb1b\") " Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.388435 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:28:51 crc kubenswrapper[4809]: E1205 12:28:51.388868 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerName="rabbitmq" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.388888 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerName="rabbitmq" Dec 05 12:28:51 crc kubenswrapper[4809]: E1205 12:28:51.388898 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerName="setup-container" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.388904 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerName="setup-container" Dec 05 12:28:51 crc kubenswrapper[4809]: E1205 12:28:51.388932 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerName="dnsmasq-dns" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.388938 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerName="dnsmasq-dns" Dec 05 12:28:51 crc kubenswrapper[4809]: E1205 12:28:51.388953 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerName="init" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.388959 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerName="init" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.389108 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" containerName="dnsmasq-dns" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.389131 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" containerName="rabbitmq" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.390103 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.391844 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.392100 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.392989 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.393226 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-t9ch6" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.393342 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.405379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-kube-api-access-xzvsw" (OuterVolumeSpecName: "kube-api-access-xzvsw") pod "fde5555e-3d40-4b92-ab47-f9b2043ddb1b" (UID: "fde5555e-3d40-4b92-ab47-f9b2043ddb1b"). InnerVolumeSpecName "kube-api-access-xzvsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.408851 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.436538 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-config" (OuterVolumeSpecName: "config") pod "fde5555e-3d40-4b92-ab47-f9b2043ddb1b" (UID: "fde5555e-3d40-4b92-ab47-f9b2043ddb1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.436985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fde5555e-3d40-4b92-ab47-f9b2043ddb1b" (UID: "fde5555e-3d40-4b92-ab47-f9b2043ddb1b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.483981 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzvsw\" (UniqueName: \"kubernetes.io/projected/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-kube-api-access-xzvsw\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.484009 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.484020 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde5555e-3d40-4b92-ab47-f9b2043ddb1b-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.585889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.585953 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a621b075-5a72-4a2a-992f-a927ffb98a57-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.585991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a621b075-5a72-4a2a-992f-a927ffb98a57-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.586018 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a621b075-5a72-4a2a-992f-a927ffb98a57-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.586059 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.586114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.586196 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.586233 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a621b075-5a72-4a2a-992f-a927ffb98a57-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.586274 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgzj7\" (UniqueName: \"kubernetes.io/projected/a621b075-5a72-4a2a-992f-a927ffb98a57-kube-api-access-cgzj7\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a621b075-5a72-4a2a-992f-a927ffb98a57-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688393 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgzj7\" (UniqueName: \"kubernetes.io/projected/a621b075-5a72-4a2a-992f-a927ffb98a57-kube-api-access-cgzj7\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688462 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688512 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a621b075-5a72-4a2a-992f-a927ffb98a57-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688565 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a621b075-5a72-4a2a-992f-a927ffb98a57-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688618 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a621b075-5a72-4a2a-992f-a927ffb98a57-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.688720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.689231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.689261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.689867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a621b075-5a72-4a2a-992f-a927ffb98a57-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.690362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a621b075-5a72-4a2a-992f-a927ffb98a57-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.691781 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.691841 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cef7a9c2f28e824d64b9cf8b9bdb2eb2a26949218d41e5db25547eafe60575a4/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.692794 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a621b075-5a72-4a2a-992f-a927ffb98a57-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.705143 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a621b075-5a72-4a2a-992f-a927ffb98a57-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.706078 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a621b075-5a72-4a2a-992f-a927ffb98a57-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.708809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgzj7\" (UniqueName: \"kubernetes.io/projected/a621b075-5a72-4a2a-992f-a927ffb98a57-kube-api-access-cgzj7\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.747941 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c24c3a-551a-4b0f-811e-5399cbdfd721\") pod \"rabbitmq-cell1-server-0\" (UID: \"a621b075-5a72-4a2a-992f-a927ffb98a57\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:51 crc kubenswrapper[4809]: I1205 12:28:51.783923 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.048560 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.048547 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-9gdsv" event={"ID":"fde5555e-3d40-4b92-ab47-f9b2043ddb1b","Type":"ContainerDied","Data":"970d59866285e5608cb3f78b8d3ba8d195fbecc16987f81e5b8f41a11859ab14"} Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.048712 4809 scope.go:117] "RemoveContainer" containerID="adc6b1eedf990527e5509a77d0678ce0582266094a8fb061799998318c6d30f0" Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.083478 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-9gdsv"] Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.085966 4809 scope.go:117] "RemoveContainer" containerID="dd45f23e4bc76c0e1e489627c0bc7e8d3b39515b78dcd5fae57c662667a8dafb" Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.093660 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-9gdsv"] Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.284518 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 12:28:52 crc kubenswrapper[4809]: W1205 12:28:52.288148 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda621b075_5a72_4a2a_992f_a927ffb98a57.slice/crio-45fcbed4dcf5fa737a3d25bee04e06945146e3f377171835a77fdc0d767c5633 WatchSource:0}: Error finding container 45fcbed4dcf5fa737a3d25bee04e06945146e3f377171835a77fdc0d767c5633: Status 404 returned error can't find the container with id 45fcbed4dcf5fa737a3d25bee04e06945146e3f377171835a77fdc0d767c5633 Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.884987 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87" path="/var/lib/kubelet/pods/ee55ec3a-b3e8-4d18-bac6-4dc66ddabb87/volumes" Dec 05 12:28:52 crc kubenswrapper[4809]: I1205 12:28:52.885866 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fde5555e-3d40-4b92-ab47-f9b2043ddb1b" path="/var/lib/kubelet/pods/fde5555e-3d40-4b92-ab47-f9b2043ddb1b/volumes" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.018565 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6z9fk"] Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.022365 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.033988 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6z9fk"] Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.071509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5b994bc1-9a84-41c6-97df-6ee239d92968","Type":"ContainerStarted","Data":"3fc9e80d58bf6bc22cf6ac1d112c81d4b5a7f9b0af9300e8fcebdbdee4027b68"} Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.076359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a621b075-5a72-4a2a-992f-a927ffb98a57","Type":"ContainerStarted","Data":"45fcbed4dcf5fa737a3d25bee04e06945146e3f377171835a77fdc0d767c5633"} Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.212121 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-utilities\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.212257 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg42s\" (UniqueName: \"kubernetes.io/projected/5f103aeb-ae36-4985-b398-fee8334f5915-kube-api-access-pg42s\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.212708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-catalog-content\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.314326 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-utilities\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.314414 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg42s\" (UniqueName: \"kubernetes.io/projected/5f103aeb-ae36-4985-b398-fee8334f5915-kube-api-access-pg42s\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.314486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-catalog-content\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.314984 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-catalog-content\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.315228 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-utilities\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.338588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg42s\" (UniqueName: \"kubernetes.io/projected/5f103aeb-ae36-4985-b398-fee8334f5915-kube-api-access-pg42s\") pod \"redhat-operators-6z9fk\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.359125 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:28:53 crc kubenswrapper[4809]: W1205 12:28:53.801290 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f103aeb_ae36_4985_b398_fee8334f5915.slice/crio-d80922e6c9489d350d97fa98a33ee14d5f3f46faaa0330b58a2c412d22ce850a WatchSource:0}: Error finding container d80922e6c9489d350d97fa98a33ee14d5f3f46faaa0330b58a2c412d22ce850a: Status 404 returned error can't find the container with id d80922e6c9489d350d97fa98a33ee14d5f3f46faaa0330b58a2c412d22ce850a Dec 05 12:28:53 crc kubenswrapper[4809]: I1205 12:28:53.803465 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6z9fk"] Dec 05 12:28:54 crc kubenswrapper[4809]: I1205 12:28:54.083343 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a621b075-5a72-4a2a-992f-a927ffb98a57","Type":"ContainerStarted","Data":"f6b8a213874b1d13e4e7d1fad53dc33acb2bb690db96511cd3e06a544399010d"} Dec 05 12:28:54 crc kubenswrapper[4809]: I1205 12:28:54.086383 4809 generic.go:334] "Generic (PLEG): container finished" podID="5f103aeb-ae36-4985-b398-fee8334f5915" containerID="543bf3236032ff17d766ee794829faabc2f71e67a3b12af4b6150fa7bed7b38d" exitCode=0 Dec 05 12:28:54 crc kubenswrapper[4809]: I1205 12:28:54.086500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9fk" event={"ID":"5f103aeb-ae36-4985-b398-fee8334f5915","Type":"ContainerDied","Data":"543bf3236032ff17d766ee794829faabc2f71e67a3b12af4b6150fa7bed7b38d"} Dec 05 12:28:54 crc kubenswrapper[4809]: I1205 12:28:54.086600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9fk" event={"ID":"5f103aeb-ae36-4985-b398-fee8334f5915","Type":"ContainerStarted","Data":"d80922e6c9489d350d97fa98a33ee14d5f3f46faaa0330b58a2c412d22ce850a"} Dec 05 12:28:55 crc kubenswrapper[4809]: I1205 12:28:55.095824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9fk" event={"ID":"5f103aeb-ae36-4985-b398-fee8334f5915","Type":"ContainerStarted","Data":"cde500cd036845c9279c90d30963ba84f3381915ed6cea087236c1e33250e9e1"} Dec 05 12:28:56 crc kubenswrapper[4809]: I1205 12:28:56.105927 4809 generic.go:334] "Generic (PLEG): container finished" podID="5f103aeb-ae36-4985-b398-fee8334f5915" containerID="cde500cd036845c9279c90d30963ba84f3381915ed6cea087236c1e33250e9e1" exitCode=0 Dec 05 12:28:56 crc kubenswrapper[4809]: I1205 12:28:56.106028 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9fk" event={"ID":"5f103aeb-ae36-4985-b398-fee8334f5915","Type":"ContainerDied","Data":"cde500cd036845c9279c90d30963ba84f3381915ed6cea087236c1e33250e9e1"} Dec 05 12:28:57 crc kubenswrapper[4809]: I1205 12:28:57.117379 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9fk" event={"ID":"5f103aeb-ae36-4985-b398-fee8334f5915","Type":"ContainerStarted","Data":"20bbcf880716992bb10139488558170ed53aa699fdac3671bcd9a823da27c80b"} Dec 05 12:28:57 crc kubenswrapper[4809]: I1205 12:28:57.138172 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6z9fk" podStartSLOduration=2.6100245319999997 podStartE2EDuration="5.138154638s" podCreationTimestamp="2025-12-05 12:28:52 +0000 UTC" firstStartedPulling="2025-12-05 12:28:54.087709781 +0000 UTC m=+4829.478686339" lastFinishedPulling="2025-12-05 12:28:56.615839867 +0000 UTC m=+4832.006816445" observedRunningTime="2025-12-05 12:28:57.136773591 +0000 UTC m=+4832.527750179" watchObservedRunningTime="2025-12-05 12:28:57.138154638 +0000 UTC m=+4832.529131196" Dec 05 12:29:03 crc kubenswrapper[4809]: I1205 12:29:03.359500 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:29:03 crc kubenswrapper[4809]: I1205 12:29:03.360334 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:29:03 crc kubenswrapper[4809]: I1205 12:29:03.411894 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:29:04 crc kubenswrapper[4809]: I1205 12:29:04.244927 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:29:04 crc kubenswrapper[4809]: I1205 12:29:04.294515 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6z9fk"] Dec 05 12:29:06 crc kubenswrapper[4809]: I1205 12:29:06.195723 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6z9fk" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="registry-server" containerID="cri-o://20bbcf880716992bb10139488558170ed53aa699fdac3671bcd9a823da27c80b" gracePeriod=2 Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.212389 4809 generic.go:334] "Generic (PLEG): container finished" podID="5f103aeb-ae36-4985-b398-fee8334f5915" containerID="20bbcf880716992bb10139488558170ed53aa699fdac3671bcd9a823da27c80b" exitCode=0 Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.212471 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9fk" event={"ID":"5f103aeb-ae36-4985-b398-fee8334f5915","Type":"ContainerDied","Data":"20bbcf880716992bb10139488558170ed53aa699fdac3671bcd9a823da27c80b"} Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.575761 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.672933 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-catalog-content\") pod \"5f103aeb-ae36-4985-b398-fee8334f5915\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.673056 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-utilities\") pod \"5f103aeb-ae36-4985-b398-fee8334f5915\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.673098 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg42s\" (UniqueName: \"kubernetes.io/projected/5f103aeb-ae36-4985-b398-fee8334f5915-kube-api-access-pg42s\") pod \"5f103aeb-ae36-4985-b398-fee8334f5915\" (UID: \"5f103aeb-ae36-4985-b398-fee8334f5915\") " Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.674118 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-utilities" (OuterVolumeSpecName: "utilities") pod "5f103aeb-ae36-4985-b398-fee8334f5915" (UID: "5f103aeb-ae36-4985-b398-fee8334f5915"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.678536 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f103aeb-ae36-4985-b398-fee8334f5915-kube-api-access-pg42s" (OuterVolumeSpecName: "kube-api-access-pg42s") pod "5f103aeb-ae36-4985-b398-fee8334f5915" (UID: "5f103aeb-ae36-4985-b398-fee8334f5915"). InnerVolumeSpecName "kube-api-access-pg42s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.775794 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.775881 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg42s\" (UniqueName: \"kubernetes.io/projected/5f103aeb-ae36-4985-b398-fee8334f5915-kube-api-access-pg42s\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.786324 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f103aeb-ae36-4985-b398-fee8334f5915" (UID: "5f103aeb-ae36-4985-b398-fee8334f5915"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:29:08 crc kubenswrapper[4809]: I1205 12:29:08.877701 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f103aeb-ae36-4985-b398-fee8334f5915-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:09 crc kubenswrapper[4809]: I1205 12:29:09.221154 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9fk" event={"ID":"5f103aeb-ae36-4985-b398-fee8334f5915","Type":"ContainerDied","Data":"d80922e6c9489d350d97fa98a33ee14d5f3f46faaa0330b58a2c412d22ce850a"} Dec 05 12:29:09 crc kubenswrapper[4809]: I1205 12:29:09.221210 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9fk" Dec 05 12:29:09 crc kubenswrapper[4809]: I1205 12:29:09.221220 4809 scope.go:117] "RemoveContainer" containerID="20bbcf880716992bb10139488558170ed53aa699fdac3671bcd9a823da27c80b" Dec 05 12:29:09 crc kubenswrapper[4809]: I1205 12:29:09.247548 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6z9fk"] Dec 05 12:29:09 crc kubenswrapper[4809]: I1205 12:29:09.253865 4809 scope.go:117] "RemoveContainer" containerID="cde500cd036845c9279c90d30963ba84f3381915ed6cea087236c1e33250e9e1" Dec 05 12:29:09 crc kubenswrapper[4809]: I1205 12:29:09.254697 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6z9fk"] Dec 05 12:29:09 crc kubenswrapper[4809]: I1205 12:29:09.273365 4809 scope.go:117] "RemoveContainer" containerID="543bf3236032ff17d766ee794829faabc2f71e67a3b12af4b6150fa7bed7b38d" Dec 05 12:29:10 crc kubenswrapper[4809]: I1205 12:29:10.882069 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" path="/var/lib/kubelet/pods/5f103aeb-ae36-4985-b398-fee8334f5915/volumes" Dec 05 12:29:25 crc kubenswrapper[4809]: I1205 12:29:25.348151 4809 generic.go:334] "Generic (PLEG): container finished" podID="5b994bc1-9a84-41c6-97df-6ee239d92968" containerID="3fc9e80d58bf6bc22cf6ac1d112c81d4b5a7f9b0af9300e8fcebdbdee4027b68" exitCode=0 Dec 05 12:29:25 crc kubenswrapper[4809]: I1205 12:29:25.348243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5b994bc1-9a84-41c6-97df-6ee239d92968","Type":"ContainerDied","Data":"3fc9e80d58bf6bc22cf6ac1d112c81d4b5a7f9b0af9300e8fcebdbdee4027b68"} Dec 05 12:29:26 crc kubenswrapper[4809]: I1205 12:29:26.360314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5b994bc1-9a84-41c6-97df-6ee239d92968","Type":"ContainerStarted","Data":"febf54c8d573c36a4540a7a1698761a29a181f13091c2539f66aaaa52fda24af"} Dec 05 12:29:26 crc kubenswrapper[4809]: I1205 12:29:26.361492 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 12:29:26 crc kubenswrapper[4809]: I1205 12:29:26.363685 4809 generic.go:334] "Generic (PLEG): container finished" podID="a621b075-5a72-4a2a-992f-a927ffb98a57" containerID="f6b8a213874b1d13e4e7d1fad53dc33acb2bb690db96511cd3e06a544399010d" exitCode=0 Dec 05 12:29:26 crc kubenswrapper[4809]: I1205 12:29:26.363748 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a621b075-5a72-4a2a-992f-a927ffb98a57","Type":"ContainerDied","Data":"f6b8a213874b1d13e4e7d1fad53dc33acb2bb690db96511cd3e06a544399010d"} Dec 05 12:29:26 crc kubenswrapper[4809]: I1205 12:29:26.393061 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.393037752 podStartE2EDuration="36.393037752s" podCreationTimestamp="2025-12-05 12:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:29:26.384095459 +0000 UTC m=+4861.775072037" watchObservedRunningTime="2025-12-05 12:29:26.393037752 +0000 UTC m=+4861.784014310" Dec 05 12:29:27 crc kubenswrapper[4809]: I1205 12:29:27.371216 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a621b075-5a72-4a2a-992f-a927ffb98a57","Type":"ContainerStarted","Data":"bb6608bdcf23af89c028f6edc75dd557372c7b93536a75c57d75e3b1ee69f5d9"} Dec 05 12:29:27 crc kubenswrapper[4809]: I1205 12:29:27.371774 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:29:27 crc kubenswrapper[4809]: I1205 12:29:27.400443 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.400424471 podStartE2EDuration="36.400424471s" podCreationTimestamp="2025-12-05 12:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:29:27.393792631 +0000 UTC m=+4862.784769199" watchObservedRunningTime="2025-12-05 12:29:27.400424471 +0000 UTC m=+4862.791401029" Dec 05 12:29:40 crc kubenswrapper[4809]: I1205 12:29:40.456839 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 12:29:41 crc kubenswrapper[4809]: I1205 12:29:41.786814 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.447438 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 12:29:52 crc kubenswrapper[4809]: E1205 12:29:52.448706 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="extract-content" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.448727 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="extract-content" Dec 05 12:29:52 crc kubenswrapper[4809]: E1205 12:29:52.448743 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="registry-server" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.448752 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="registry-server" Dec 05 12:29:52 crc kubenswrapper[4809]: E1205 12:29:52.448780 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="extract-utilities" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.448813 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="extract-utilities" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.448991 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f103aeb-ae36-4985-b398-fee8334f5915" containerName="registry-server" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.449559 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.451986 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kdqm4" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.456916 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.574114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcrlr\" (UniqueName: \"kubernetes.io/projected/e9600fd9-b5a3-4ff4-826c-85308b6a7b72-kube-api-access-mcrlr\") pod \"mariadb-client-1-default\" (UID: \"e9600fd9-b5a3-4ff4-826c-85308b6a7b72\") " pod="openstack/mariadb-client-1-default" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.675497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcrlr\" (UniqueName: \"kubernetes.io/projected/e9600fd9-b5a3-4ff4-826c-85308b6a7b72-kube-api-access-mcrlr\") pod \"mariadb-client-1-default\" (UID: \"e9600fd9-b5a3-4ff4-826c-85308b6a7b72\") " pod="openstack/mariadb-client-1-default" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.696358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcrlr\" (UniqueName: \"kubernetes.io/projected/e9600fd9-b5a3-4ff4-826c-85308b6a7b72-kube-api-access-mcrlr\") pod \"mariadb-client-1-default\" (UID: \"e9600fd9-b5a3-4ff4-826c-85308b6a7b72\") " pod="openstack/mariadb-client-1-default" Dec 05 12:29:52 crc kubenswrapper[4809]: I1205 12:29:52.773862 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 12:29:53 crc kubenswrapper[4809]: I1205 12:29:53.343581 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 12:29:53 crc kubenswrapper[4809]: W1205 12:29:53.353727 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9600fd9_b5a3_4ff4_826c_85308b6a7b72.slice/crio-0867d4e06a50488bcffea1704e1a8afb096f84732d8bb33fffcd5d48cf02598e WatchSource:0}: Error finding container 0867d4e06a50488bcffea1704e1a8afb096f84732d8bb33fffcd5d48cf02598e: Status 404 returned error can't find the container with id 0867d4e06a50488bcffea1704e1a8afb096f84732d8bb33fffcd5d48cf02598e Dec 05 12:29:53 crc kubenswrapper[4809]: I1205 12:29:53.610603 4809 generic.go:334] "Generic (PLEG): container finished" podID="e9600fd9-b5a3-4ff4-826c-85308b6a7b72" containerID="6c8088d071b24c8b405fc5952801939b6999789621074353267fd2108a946b47" exitCode=0 Dec 05 12:29:53 crc kubenswrapper[4809]: I1205 12:29:53.610671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"e9600fd9-b5a3-4ff4-826c-85308b6a7b72","Type":"ContainerDied","Data":"6c8088d071b24c8b405fc5952801939b6999789621074353267fd2108a946b47"} Dec 05 12:29:53 crc kubenswrapper[4809]: I1205 12:29:53.610702 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"e9600fd9-b5a3-4ff4-826c-85308b6a7b72","Type":"ContainerStarted","Data":"0867d4e06a50488bcffea1704e1a8afb096f84732d8bb33fffcd5d48cf02598e"} Dec 05 12:29:54 crc kubenswrapper[4809]: I1205 12:29:54.977180 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.001867 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_e9600fd9-b5a3-4ff4-826c-85308b6a7b72/mariadb-client-1-default/0.log" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.024663 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.030690 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.111783 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcrlr\" (UniqueName: \"kubernetes.io/projected/e9600fd9-b5a3-4ff4-826c-85308b6a7b72-kube-api-access-mcrlr\") pod \"e9600fd9-b5a3-4ff4-826c-85308b6a7b72\" (UID: \"e9600fd9-b5a3-4ff4-826c-85308b6a7b72\") " Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.119166 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9600fd9-b5a3-4ff4-826c-85308b6a7b72-kube-api-access-mcrlr" (OuterVolumeSpecName: "kube-api-access-mcrlr") pod "e9600fd9-b5a3-4ff4-826c-85308b6a7b72" (UID: "e9600fd9-b5a3-4ff4-826c-85308b6a7b72"). InnerVolumeSpecName "kube-api-access-mcrlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.213306 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcrlr\" (UniqueName: \"kubernetes.io/projected/e9600fd9-b5a3-4ff4-826c-85308b6a7b72-kube-api-access-mcrlr\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.437483 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 12:29:55 crc kubenswrapper[4809]: E1205 12:29:55.438200 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9600fd9-b5a3-4ff4-826c-85308b6a7b72" containerName="mariadb-client-1-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.438226 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9600fd9-b5a3-4ff4-826c-85308b6a7b72" containerName="mariadb-client-1-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.438446 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9600fd9-b5a3-4ff4-826c-85308b6a7b72" containerName="mariadb-client-1-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.439127 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.456968 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.618624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q4gq\" (UniqueName: \"kubernetes.io/projected/7116ecc2-9fd9-40ef-9520-7429a213fe63-kube-api-access-8q4gq\") pod \"mariadb-client-2-default\" (UID: \"7116ecc2-9fd9-40ef-9520-7429a213fe63\") " pod="openstack/mariadb-client-2-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.625677 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0867d4e06a50488bcffea1704e1a8afb096f84732d8bb33fffcd5d48cf02598e" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.625781 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.720615 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q4gq\" (UniqueName: \"kubernetes.io/projected/7116ecc2-9fd9-40ef-9520-7429a213fe63-kube-api-access-8q4gq\") pod \"mariadb-client-2-default\" (UID: \"7116ecc2-9fd9-40ef-9520-7429a213fe63\") " pod="openstack/mariadb-client-2-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.744507 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q4gq\" (UniqueName: \"kubernetes.io/projected/7116ecc2-9fd9-40ef-9520-7429a213fe63-kube-api-access-8q4gq\") pod \"mariadb-client-2-default\" (UID: \"7116ecc2-9fd9-40ef-9520-7429a213fe63\") " pod="openstack/mariadb-client-2-default" Dec 05 12:29:55 crc kubenswrapper[4809]: I1205 12:29:55.763775 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 12:29:56 crc kubenswrapper[4809]: I1205 12:29:56.370405 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 12:29:56 crc kubenswrapper[4809]: W1205 12:29:56.371523 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7116ecc2_9fd9_40ef_9520_7429a213fe63.slice/crio-37c1742c5b52b685d00595d856717d5d2599c7dd330e33d756987266307c4291 WatchSource:0}: Error finding container 37c1742c5b52b685d00595d856717d5d2599c7dd330e33d756987266307c4291: Status 404 returned error can't find the container with id 37c1742c5b52b685d00595d856717d5d2599c7dd330e33d756987266307c4291 Dec 05 12:29:56 crc kubenswrapper[4809]: I1205 12:29:56.634423 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"7116ecc2-9fd9-40ef-9520-7429a213fe63","Type":"ContainerStarted","Data":"ed6561da390538c84db140498382635d42afa5a4707dbda0a96a39452d28f77a"} Dec 05 12:29:56 crc kubenswrapper[4809]: I1205 12:29:56.634917 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"7116ecc2-9fd9-40ef-9520-7429a213fe63","Type":"ContainerStarted","Data":"37c1742c5b52b685d00595d856717d5d2599c7dd330e33d756987266307c4291"} Dec 05 12:29:56 crc kubenswrapper[4809]: I1205 12:29:56.653424 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.6534075910000001 podStartE2EDuration="1.653407591s" podCreationTimestamp="2025-12-05 12:29:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:29:56.647416419 +0000 UTC m=+4892.038392977" watchObservedRunningTime="2025-12-05 12:29:56.653407591 +0000 UTC m=+4892.044384149" Dec 05 12:29:56 crc kubenswrapper[4809]: I1205 12:29:56.885033 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9600fd9-b5a3-4ff4-826c-85308b6a7b72" path="/var/lib/kubelet/pods/e9600fd9-b5a3-4ff4-826c-85308b6a7b72/volumes" Dec 05 12:29:57 crc kubenswrapper[4809]: I1205 12:29:57.650165 4809 generic.go:334] "Generic (PLEG): container finished" podID="7116ecc2-9fd9-40ef-9520-7429a213fe63" containerID="ed6561da390538c84db140498382635d42afa5a4707dbda0a96a39452d28f77a" exitCode=1 Dec 05 12:29:57 crc kubenswrapper[4809]: I1205 12:29:57.651578 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"7116ecc2-9fd9-40ef-9520-7429a213fe63","Type":"ContainerDied","Data":"ed6561da390538c84db140498382635d42afa5a4707dbda0a96a39452d28f77a"} Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.039375 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.077577 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.085895 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.174920 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q4gq\" (UniqueName: \"kubernetes.io/projected/7116ecc2-9fd9-40ef-9520-7429a213fe63-kube-api-access-8q4gq\") pod \"7116ecc2-9fd9-40ef-9520-7429a213fe63\" (UID: \"7116ecc2-9fd9-40ef-9520-7429a213fe63\") " Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.180041 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7116ecc2-9fd9-40ef-9520-7429a213fe63-kube-api-access-8q4gq" (OuterVolumeSpecName: "kube-api-access-8q4gq") pod "7116ecc2-9fd9-40ef-9520-7429a213fe63" (UID: "7116ecc2-9fd9-40ef-9520-7429a213fe63"). InnerVolumeSpecName "kube-api-access-8q4gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.276880 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q4gq\" (UniqueName: \"kubernetes.io/projected/7116ecc2-9fd9-40ef-9520-7429a213fe63-kube-api-access-8q4gq\") on node \"crc\" DevicePath \"\"" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.491621 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 05 12:29:59 crc kubenswrapper[4809]: E1205 12:29:59.492049 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7116ecc2-9fd9-40ef-9520-7429a213fe63" containerName="mariadb-client-2-default" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.492072 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7116ecc2-9fd9-40ef-9520-7429a213fe63" containerName="mariadb-client-2-default" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.492318 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7116ecc2-9fd9-40ef-9520-7429a213fe63" containerName="mariadb-client-2-default" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.492987 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.509579 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.582215 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5cnk\" (UniqueName: \"kubernetes.io/projected/ddcda4c5-7dce-45ff-bffc-5ac57831bcb2-kube-api-access-j5cnk\") pod \"mariadb-client-1\" (UID: \"ddcda4c5-7dce-45ff-bffc-5ac57831bcb2\") " pod="openstack/mariadb-client-1" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.667150 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37c1742c5b52b685d00595d856717d5d2599c7dd330e33d756987266307c4291" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.667242 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.683412 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5cnk\" (UniqueName: \"kubernetes.io/projected/ddcda4c5-7dce-45ff-bffc-5ac57831bcb2-kube-api-access-j5cnk\") pod \"mariadb-client-1\" (UID: \"ddcda4c5-7dce-45ff-bffc-5ac57831bcb2\") " pod="openstack/mariadb-client-1" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.701263 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5cnk\" (UniqueName: \"kubernetes.io/projected/ddcda4c5-7dce-45ff-bffc-5ac57831bcb2-kube-api-access-j5cnk\") pod \"mariadb-client-1\" (UID: \"ddcda4c5-7dce-45ff-bffc-5ac57831bcb2\") " pod="openstack/mariadb-client-1" Dec 05 12:29:59 crc kubenswrapper[4809]: I1205 12:29:59.821373 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.146365 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5"] Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.149465 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.152436 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.152871 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.157678 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5"] Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.292519 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/130471fd-53bd-4db9-9ba0-76e55f3778a2-secret-volume\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.292821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2mtd\" (UniqueName: \"kubernetes.io/projected/130471fd-53bd-4db9-9ba0-76e55f3778a2-kube-api-access-h2mtd\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.293064 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/130471fd-53bd-4db9-9ba0-76e55f3778a2-config-volume\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.394502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/130471fd-53bd-4db9-9ba0-76e55f3778a2-secret-volume\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.394594 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2mtd\" (UniqueName: \"kubernetes.io/projected/130471fd-53bd-4db9-9ba0-76e55f3778a2-kube-api-access-h2mtd\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.394709 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/130471fd-53bd-4db9-9ba0-76e55f3778a2-config-volume\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.395915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/130471fd-53bd-4db9-9ba0-76e55f3778a2-config-volume\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.401910 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/130471fd-53bd-4db9-9ba0-76e55f3778a2-secret-volume\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.411917 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2mtd\" (UniqueName: \"kubernetes.io/projected/130471fd-53bd-4db9-9ba0-76e55f3778a2-kube-api-access-h2mtd\") pod \"collect-profiles-29415630-bh8n5\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.474825 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:00 crc kubenswrapper[4809]: I1205 12:30:00.880743 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7116ecc2-9fd9-40ef-9520-7429a213fe63" path="/var/lib/kubelet/pods/7116ecc2-9fd9-40ef-9520-7429a213fe63/volumes" Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.153514 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5"] Dec 05 12:30:01 crc kubenswrapper[4809]: W1205 12:30:01.173835 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod130471fd_53bd_4db9_9ba0_76e55f3778a2.slice/crio-ffc0a37655bf456065d7768695fafa308c4c6075e7b1d63143ea7a3118b1c28b WatchSource:0}: Error finding container ffc0a37655bf456065d7768695fafa308c4c6075e7b1d63143ea7a3118b1c28b: Status 404 returned error can't find the container with id ffc0a37655bf456065d7768695fafa308c4c6075e7b1d63143ea7a3118b1c28b Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.175810 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 12:30:01 crc kubenswrapper[4809]: W1205 12:30:01.197844 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddcda4c5_7dce_45ff_bffc_5ac57831bcb2.slice/crio-bfceecf665971714c4e844ff7625c9c841aa996cf9d45652ff04805d9a7d26b5 WatchSource:0}: Error finding container bfceecf665971714c4e844ff7625c9c841aa996cf9d45652ff04805d9a7d26b5: Status 404 returned error can't find the container with id bfceecf665971714c4e844ff7625c9c841aa996cf9d45652ff04805d9a7d26b5 Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.682172 4809 generic.go:334] "Generic (PLEG): container finished" podID="ddcda4c5-7dce-45ff-bffc-5ac57831bcb2" containerID="bd72a84c654487405ca915cbaccafb125378031b46d40597ae4adc95e72777fb" exitCode=0 Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.682258 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"ddcda4c5-7dce-45ff-bffc-5ac57831bcb2","Type":"ContainerDied","Data":"bd72a84c654487405ca915cbaccafb125378031b46d40597ae4adc95e72777fb"} Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.682324 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"ddcda4c5-7dce-45ff-bffc-5ac57831bcb2","Type":"ContainerStarted","Data":"bfceecf665971714c4e844ff7625c9c841aa996cf9d45652ff04805d9a7d26b5"} Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.684743 4809 generic.go:334] "Generic (PLEG): container finished" podID="130471fd-53bd-4db9-9ba0-76e55f3778a2" containerID="12b41af7869365eb8f34cb87c168476d6ac82f5d8b614b8ad6a94b91969e2931" exitCode=0 Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.684777 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" event={"ID":"130471fd-53bd-4db9-9ba0-76e55f3778a2","Type":"ContainerDied","Data":"12b41af7869365eb8f34cb87c168476d6ac82f5d8b614b8ad6a94b91969e2931"} Dec 05 12:30:01 crc kubenswrapper[4809]: I1205 12:30:01.684798 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" event={"ID":"130471fd-53bd-4db9-9ba0-76e55f3778a2","Type":"ContainerStarted","Data":"ffc0a37655bf456065d7768695fafa308c4c6075e7b1d63143ea7a3118b1c28b"} Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.108424 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.117390 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.154285 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_ddcda4c5-7dce-45ff-bffc-5ac57831bcb2/mariadb-client-1/0.log" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.239386 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.241584 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/130471fd-53bd-4db9-9ba0-76e55f3778a2-secret-volume\") pod \"130471fd-53bd-4db9-9ba0-76e55f3778a2\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.241713 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/130471fd-53bd-4db9-9ba0-76e55f3778a2-config-volume\") pod \"130471fd-53bd-4db9-9ba0-76e55f3778a2\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.241773 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2mtd\" (UniqueName: \"kubernetes.io/projected/130471fd-53bd-4db9-9ba0-76e55f3778a2-kube-api-access-h2mtd\") pod \"130471fd-53bd-4db9-9ba0-76e55f3778a2\" (UID: \"130471fd-53bd-4db9-9ba0-76e55f3778a2\") " Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.241849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5cnk\" (UniqueName: \"kubernetes.io/projected/ddcda4c5-7dce-45ff-bffc-5ac57831bcb2-kube-api-access-j5cnk\") pod \"ddcda4c5-7dce-45ff-bffc-5ac57831bcb2\" (UID: \"ddcda4c5-7dce-45ff-bffc-5ac57831bcb2\") " Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.243136 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130471fd-53bd-4db9-9ba0-76e55f3778a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "130471fd-53bd-4db9-9ba0-76e55f3778a2" (UID: "130471fd-53bd-4db9-9ba0-76e55f3778a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.248884 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130471fd-53bd-4db9-9ba0-76e55f3778a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "130471fd-53bd-4db9-9ba0-76e55f3778a2" (UID: "130471fd-53bd-4db9-9ba0-76e55f3778a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.249038 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130471fd-53bd-4db9-9ba0-76e55f3778a2-kube-api-access-h2mtd" (OuterVolumeSpecName: "kube-api-access-h2mtd") pod "130471fd-53bd-4db9-9ba0-76e55f3778a2" (UID: "130471fd-53bd-4db9-9ba0-76e55f3778a2"). InnerVolumeSpecName "kube-api-access-h2mtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.250098 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddcda4c5-7dce-45ff-bffc-5ac57831bcb2-kube-api-access-j5cnk" (OuterVolumeSpecName: "kube-api-access-j5cnk") pod "ddcda4c5-7dce-45ff-bffc-5ac57831bcb2" (UID: "ddcda4c5-7dce-45ff-bffc-5ac57831bcb2"). InnerVolumeSpecName "kube-api-access-j5cnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.250260 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.344040 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/130471fd-53bd-4db9-9ba0-76e55f3778a2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.344392 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2mtd\" (UniqueName: \"kubernetes.io/projected/130471fd-53bd-4db9-9ba0-76e55f3778a2-kube-api-access-h2mtd\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.344406 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5cnk\" (UniqueName: \"kubernetes.io/projected/ddcda4c5-7dce-45ff-bffc-5ac57831bcb2-kube-api-access-j5cnk\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.344418 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/130471fd-53bd-4db9-9ba0-76e55f3778a2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.642348 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 12:30:03 crc kubenswrapper[4809]: E1205 12:30:03.642697 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130471fd-53bd-4db9-9ba0-76e55f3778a2" containerName="collect-profiles" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.642723 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="130471fd-53bd-4db9-9ba0-76e55f3778a2" containerName="collect-profiles" Dec 05 12:30:03 crc kubenswrapper[4809]: E1205 12:30:03.642743 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddcda4c5-7dce-45ff-bffc-5ac57831bcb2" containerName="mariadb-client-1" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.642750 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddcda4c5-7dce-45ff-bffc-5ac57831bcb2" containerName="mariadb-client-1" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.642901 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="130471fd-53bd-4db9-9ba0-76e55f3778a2" containerName="collect-profiles" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.642918 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddcda4c5-7dce-45ff-bffc-5ac57831bcb2" containerName="mariadb-client-1" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.643431 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.655524 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.706494 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" event={"ID":"130471fd-53bd-4db9-9ba0-76e55f3778a2","Type":"ContainerDied","Data":"ffc0a37655bf456065d7768695fafa308c4c6075e7b1d63143ea7a3118b1c28b"} Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.706541 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.706554 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffc0a37655bf456065d7768695fafa308c4c6075e7b1d63143ea7a3118b1c28b" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.708134 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.708137 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfceecf665971714c4e844ff7625c9c841aa996cf9d45652ff04805d9a7d26b5" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.750027 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlvnx\" (UniqueName: \"kubernetes.io/projected/6bb89dc8-78d5-42b8-8b53-b2692ff6a080-kube-api-access-wlvnx\") pod \"mariadb-client-4-default\" (UID: \"6bb89dc8-78d5-42b8-8b53-b2692ff6a080\") " pod="openstack/mariadb-client-4-default" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.852173 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlvnx\" (UniqueName: \"kubernetes.io/projected/6bb89dc8-78d5-42b8-8b53-b2692ff6a080-kube-api-access-wlvnx\") pod \"mariadb-client-4-default\" (UID: \"6bb89dc8-78d5-42b8-8b53-b2692ff6a080\") " pod="openstack/mariadb-client-4-default" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.869725 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlvnx\" (UniqueName: \"kubernetes.io/projected/6bb89dc8-78d5-42b8-8b53-b2692ff6a080-kube-api-access-wlvnx\") pod \"mariadb-client-4-default\" (UID: \"6bb89dc8-78d5-42b8-8b53-b2692ff6a080\") " pod="openstack/mariadb-client-4-default" Dec 05 12:30:03 crc kubenswrapper[4809]: I1205 12:30:03.975749 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.204824 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm"] Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.212119 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-j55qm"] Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.294716 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.717574 4809 generic.go:334] "Generic (PLEG): container finished" podID="6bb89dc8-78d5-42b8-8b53-b2692ff6a080" containerID="19d873abe1e7ed6f397baf9fc551509b0c5960a456d6ded51994fc9326d9c3ce" exitCode=0 Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.717615 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"6bb89dc8-78d5-42b8-8b53-b2692ff6a080","Type":"ContainerDied","Data":"19d873abe1e7ed6f397baf9fc551509b0c5960a456d6ded51994fc9326d9c3ce"} Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.717659 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"6bb89dc8-78d5-42b8-8b53-b2692ff6a080","Type":"ContainerStarted","Data":"065d8a482060b72aaf4fb1f92986a79445b9a8b4c5e64b17f48e757e5795a9f6"} Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.880934 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b9cf7e7-71b9-47df-ac05-09bc6406360f" path="/var/lib/kubelet/pods/7b9cf7e7-71b9-47df-ac05-09bc6406360f/volumes" Dec 05 12:30:04 crc kubenswrapper[4809]: I1205 12:30:04.881778 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddcda4c5-7dce-45ff-bffc-5ac57831bcb2" path="/var/lib/kubelet/pods/ddcda4c5-7dce-45ff-bffc-5ac57831bcb2/volumes" Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.088375 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.116330 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_6bb89dc8-78d5-42b8-8b53-b2692ff6a080/mariadb-client-4-default/0.log" Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.152427 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.157816 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.187722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlvnx\" (UniqueName: \"kubernetes.io/projected/6bb89dc8-78d5-42b8-8b53-b2692ff6a080-kube-api-access-wlvnx\") pod \"6bb89dc8-78d5-42b8-8b53-b2692ff6a080\" (UID: \"6bb89dc8-78d5-42b8-8b53-b2692ff6a080\") " Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.206851 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bb89dc8-78d5-42b8-8b53-b2692ff6a080-kube-api-access-wlvnx" (OuterVolumeSpecName: "kube-api-access-wlvnx") pod "6bb89dc8-78d5-42b8-8b53-b2692ff6a080" (UID: "6bb89dc8-78d5-42b8-8b53-b2692ff6a080"). InnerVolumeSpecName "kube-api-access-wlvnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.290873 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlvnx\" (UniqueName: \"kubernetes.io/projected/6bb89dc8-78d5-42b8-8b53-b2692ff6a080-kube-api-access-wlvnx\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.733827 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="065d8a482060b72aaf4fb1f92986a79445b9a8b4c5e64b17f48e757e5795a9f6" Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.733886 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 05 12:30:06 crc kubenswrapper[4809]: I1205 12:30:06.884044 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bb89dc8-78d5-42b8-8b53-b2692ff6a080" path="/var/lib/kubelet/pods/6bb89dc8-78d5-42b8-8b53-b2692ff6a080/volumes" Dec 05 12:30:09 crc kubenswrapper[4809]: I1205 12:30:09.869572 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 12:30:09 crc kubenswrapper[4809]: E1205 12:30:09.870210 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb89dc8-78d5-42b8-8b53-b2692ff6a080" containerName="mariadb-client-4-default" Dec 05 12:30:09 crc kubenswrapper[4809]: I1205 12:30:09.870222 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb89dc8-78d5-42b8-8b53-b2692ff6a080" containerName="mariadb-client-4-default" Dec 05 12:30:09 crc kubenswrapper[4809]: I1205 12:30:09.870374 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb89dc8-78d5-42b8-8b53-b2692ff6a080" containerName="mariadb-client-4-default" Dec 05 12:30:09 crc kubenswrapper[4809]: I1205 12:30:09.870974 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 12:30:09 crc kubenswrapper[4809]: I1205 12:30:09.873123 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kdqm4" Dec 05 12:30:09 crc kubenswrapper[4809]: I1205 12:30:09.883533 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 12:30:09 crc kubenswrapper[4809]: I1205 12:30:09.947476 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7ltx\" (UniqueName: \"kubernetes.io/projected/a1dbcb7f-2d83-445a-bfd2-5effcdee6691-kube-api-access-l7ltx\") pod \"mariadb-client-5-default\" (UID: \"a1dbcb7f-2d83-445a-bfd2-5effcdee6691\") " pod="openstack/mariadb-client-5-default" Dec 05 12:30:10 crc kubenswrapper[4809]: I1205 12:30:10.049733 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7ltx\" (UniqueName: \"kubernetes.io/projected/a1dbcb7f-2d83-445a-bfd2-5effcdee6691-kube-api-access-l7ltx\") pod \"mariadb-client-5-default\" (UID: \"a1dbcb7f-2d83-445a-bfd2-5effcdee6691\") " pod="openstack/mariadb-client-5-default" Dec 05 12:30:10 crc kubenswrapper[4809]: I1205 12:30:10.075101 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7ltx\" (UniqueName: \"kubernetes.io/projected/a1dbcb7f-2d83-445a-bfd2-5effcdee6691-kube-api-access-l7ltx\") pod \"mariadb-client-5-default\" (UID: \"a1dbcb7f-2d83-445a-bfd2-5effcdee6691\") " pod="openstack/mariadb-client-5-default" Dec 05 12:30:10 crc kubenswrapper[4809]: I1205 12:30:10.195229 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 12:30:10 crc kubenswrapper[4809]: I1205 12:30:10.768875 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 12:30:11 crc kubenswrapper[4809]: I1205 12:30:11.784409 4809 generic.go:334] "Generic (PLEG): container finished" podID="a1dbcb7f-2d83-445a-bfd2-5effcdee6691" containerID="c7f93ec83840866d448cf394729f17d243c69d954bc11db1a2409da09c849a3c" exitCode=0 Dec 05 12:30:11 crc kubenswrapper[4809]: I1205 12:30:11.784514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"a1dbcb7f-2d83-445a-bfd2-5effcdee6691","Type":"ContainerDied","Data":"c7f93ec83840866d448cf394729f17d243c69d954bc11db1a2409da09c849a3c"} Dec 05 12:30:11 crc kubenswrapper[4809]: I1205 12:30:11.784930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"a1dbcb7f-2d83-445a-bfd2-5effcdee6691","Type":"ContainerStarted","Data":"3f0677b576c7b7559db402fb24d77d5754756ac7c122c6fc5b917e6a170d4ead"} Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.149492 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.171476 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_a1dbcb7f-2d83-445a-bfd2-5effcdee6691/mariadb-client-5-default/0.log" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.198529 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.205903 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.211716 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7ltx\" (UniqueName: \"kubernetes.io/projected/a1dbcb7f-2d83-445a-bfd2-5effcdee6691-kube-api-access-l7ltx\") pod \"a1dbcb7f-2d83-445a-bfd2-5effcdee6691\" (UID: \"a1dbcb7f-2d83-445a-bfd2-5effcdee6691\") " Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.219326 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1dbcb7f-2d83-445a-bfd2-5effcdee6691-kube-api-access-l7ltx" (OuterVolumeSpecName: "kube-api-access-l7ltx") pod "a1dbcb7f-2d83-445a-bfd2-5effcdee6691" (UID: "a1dbcb7f-2d83-445a-bfd2-5effcdee6691"). InnerVolumeSpecName "kube-api-access-l7ltx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.313212 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7ltx\" (UniqueName: \"kubernetes.io/projected/a1dbcb7f-2d83-445a-bfd2-5effcdee6691-kube-api-access-l7ltx\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.336886 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 12:30:13 crc kubenswrapper[4809]: E1205 12:30:13.337273 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1dbcb7f-2d83-445a-bfd2-5effcdee6691" containerName="mariadb-client-5-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.337299 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1dbcb7f-2d83-445a-bfd2-5effcdee6691" containerName="mariadb-client-5-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.337490 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1dbcb7f-2d83-445a-bfd2-5effcdee6691" containerName="mariadb-client-5-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.338039 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.346099 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.516338 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm8h4\" (UniqueName: \"kubernetes.io/projected/e31b4c56-cd8a-483b-a921-a13585ebd243-kube-api-access-sm8h4\") pod \"mariadb-client-6-default\" (UID: \"e31b4c56-cd8a-483b-a921-a13585ebd243\") " pod="openstack/mariadb-client-6-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.618480 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm8h4\" (UniqueName: \"kubernetes.io/projected/e31b4c56-cd8a-483b-a921-a13585ebd243-kube-api-access-sm8h4\") pod \"mariadb-client-6-default\" (UID: \"e31b4c56-cd8a-483b-a921-a13585ebd243\") " pod="openstack/mariadb-client-6-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.649541 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm8h4\" (UniqueName: \"kubernetes.io/projected/e31b4c56-cd8a-483b-a921-a13585ebd243-kube-api-access-sm8h4\") pod \"mariadb-client-6-default\" (UID: \"e31b4c56-cd8a-483b-a921-a13585ebd243\") " pod="openstack/mariadb-client-6-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.656966 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.827177 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f0677b576c7b7559db402fb24d77d5754756ac7c122c6fc5b917e6a170d4ead" Dec 05 12:30:13 crc kubenswrapper[4809]: I1205 12:30:13.827257 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 05 12:30:14 crc kubenswrapper[4809]: I1205 12:30:14.041963 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 12:30:14 crc kubenswrapper[4809]: I1205 12:30:14.838055 4809 generic.go:334] "Generic (PLEG): container finished" podID="e31b4c56-cd8a-483b-a921-a13585ebd243" containerID="87e67d470473fa44a8d17d8f45820a07449ca12920240d23410de654f25ee331" exitCode=1 Dec 05 12:30:14 crc kubenswrapper[4809]: I1205 12:30:14.838275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"e31b4c56-cd8a-483b-a921-a13585ebd243","Type":"ContainerDied","Data":"87e67d470473fa44a8d17d8f45820a07449ca12920240d23410de654f25ee331"} Dec 05 12:30:14 crc kubenswrapper[4809]: I1205 12:30:14.838426 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"e31b4c56-cd8a-483b-a921-a13585ebd243","Type":"ContainerStarted","Data":"48c310edcd0cd4ffdfdd33b8276458055564de04ea0bdc93824d8d5c5c2a6567"} Dec 05 12:30:14 crc kubenswrapper[4809]: I1205 12:30:14.884465 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1dbcb7f-2d83-445a-bfd2-5effcdee6691" path="/var/lib/kubelet/pods/a1dbcb7f-2d83-445a-bfd2-5effcdee6691/volumes" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.191927 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.210440 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_e31b4c56-cd8a-483b-a921-a13585ebd243/mariadb-client-6-default/0.log" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.239446 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.247845 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.261257 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm8h4\" (UniqueName: \"kubernetes.io/projected/e31b4c56-cd8a-483b-a921-a13585ebd243-kube-api-access-sm8h4\") pod \"e31b4c56-cd8a-483b-a921-a13585ebd243\" (UID: \"e31b4c56-cd8a-483b-a921-a13585ebd243\") " Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.267668 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31b4c56-cd8a-483b-a921-a13585ebd243-kube-api-access-sm8h4" (OuterVolumeSpecName: "kube-api-access-sm8h4") pod "e31b4c56-cd8a-483b-a921-a13585ebd243" (UID: "e31b4c56-cd8a-483b-a921-a13585ebd243"). InnerVolumeSpecName "kube-api-access-sm8h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.362865 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm8h4\" (UniqueName: \"kubernetes.io/projected/e31b4c56-cd8a-483b-a921-a13585ebd243-kube-api-access-sm8h4\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.372873 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 12:30:16 crc kubenswrapper[4809]: E1205 12:30:16.373241 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31b4c56-cd8a-483b-a921-a13585ebd243" containerName="mariadb-client-6-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.373264 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31b4c56-cd8a-483b-a921-a13585ebd243" containerName="mariadb-client-6-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.373433 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31b4c56-cd8a-483b-a921-a13585ebd243" containerName="mariadb-client-6-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.374108 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.383105 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.565430 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkk9l\" (UniqueName: \"kubernetes.io/projected/4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8-kube-api-access-qkk9l\") pod \"mariadb-client-7-default\" (UID: \"4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8\") " pod="openstack/mariadb-client-7-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.666405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkk9l\" (UniqueName: \"kubernetes.io/projected/4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8-kube-api-access-qkk9l\") pod \"mariadb-client-7-default\" (UID: \"4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8\") " pod="openstack/mariadb-client-7-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.685992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkk9l\" (UniqueName: \"kubernetes.io/projected/4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8-kube-api-access-qkk9l\") pod \"mariadb-client-7-default\" (UID: \"4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8\") " pod="openstack/mariadb-client-7-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.703502 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.856907 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48c310edcd0cd4ffdfdd33b8276458055564de04ea0bdc93824d8d5c5c2a6567" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.856973 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.892698 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31b4c56-cd8a-483b-a921-a13585ebd243" path="/var/lib/kubelet/pods/e31b4c56-cd8a-483b-a921-a13585ebd243/volumes" Dec 05 12:30:16 crc kubenswrapper[4809]: I1205 12:30:16.959547 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 12:30:16 crc kubenswrapper[4809]: W1205 12:30:16.962955 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c1b9f1e_f3b7_41d1_bc95_29c0fc2e1ec8.slice/crio-a6187022dab19ea66189b2bd85d070dd2b5a948d2bcdec1a91876530eb36fb8a WatchSource:0}: Error finding container a6187022dab19ea66189b2bd85d070dd2b5a948d2bcdec1a91876530eb36fb8a: Status 404 returned error can't find the container with id a6187022dab19ea66189b2bd85d070dd2b5a948d2bcdec1a91876530eb36fb8a Dec 05 12:30:17 crc kubenswrapper[4809]: I1205 12:30:17.872183 4809 generic.go:334] "Generic (PLEG): container finished" podID="4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8" containerID="59ea4cd6a1691e33c5431e68b39ef81cb18ea91e85643e385c3f4ca65782d1e2" exitCode=0 Dec 05 12:30:17 crc kubenswrapper[4809]: I1205 12:30:17.872230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8","Type":"ContainerDied","Data":"59ea4cd6a1691e33c5431e68b39ef81cb18ea91e85643e385c3f4ca65782d1e2"} Dec 05 12:30:17 crc kubenswrapper[4809]: I1205 12:30:17.872256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8","Type":"ContainerStarted","Data":"a6187022dab19ea66189b2bd85d070dd2b5a948d2bcdec1a91876530eb36fb8a"} Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.233231 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.253721 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8/mariadb-client-7-default/0.log" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.280279 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.286275 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.409663 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 05 12:30:19 crc kubenswrapper[4809]: E1205 12:30:19.410647 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8" containerName="mariadb-client-7-default" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.410753 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8" containerName="mariadb-client-7-default" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.411035 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8" containerName="mariadb-client-7-default" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.411873 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.415829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkk9l\" (UniqueName: \"kubernetes.io/projected/4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8-kube-api-access-qkk9l\") pod \"4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8\" (UID: \"4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8\") " Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.418887 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.423586 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8-kube-api-access-qkk9l" (OuterVolumeSpecName: "kube-api-access-qkk9l") pod "4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8" (UID: "4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8"). InnerVolumeSpecName "kube-api-access-qkk9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.519040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvg7w\" (UniqueName: \"kubernetes.io/projected/5a0b5801-ecea-416e-a2fe-4bf8593345b3-kube-api-access-hvg7w\") pod \"mariadb-client-2\" (UID: \"5a0b5801-ecea-416e-a2fe-4bf8593345b3\") " pod="openstack/mariadb-client-2" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.519447 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkk9l\" (UniqueName: \"kubernetes.io/projected/4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8-kube-api-access-qkk9l\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.620464 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvg7w\" (UniqueName: \"kubernetes.io/projected/5a0b5801-ecea-416e-a2fe-4bf8593345b3-kube-api-access-hvg7w\") pod \"mariadb-client-2\" (UID: \"5a0b5801-ecea-416e-a2fe-4bf8593345b3\") " pod="openstack/mariadb-client-2" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.638607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvg7w\" (UniqueName: \"kubernetes.io/projected/5a0b5801-ecea-416e-a2fe-4bf8593345b3-kube-api-access-hvg7w\") pod \"mariadb-client-2\" (UID: \"5a0b5801-ecea-416e-a2fe-4bf8593345b3\") " pod="openstack/mariadb-client-2" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.763154 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.891606 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6187022dab19ea66189b2bd85d070dd2b5a948d2bcdec1a91876530eb36fb8a" Dec 05 12:30:19 crc kubenswrapper[4809]: I1205 12:30:19.891668 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 05 12:30:20 crc kubenswrapper[4809]: I1205 12:30:20.246864 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 12:30:20 crc kubenswrapper[4809]: W1205 12:30:20.252243 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a0b5801_ecea_416e_a2fe_4bf8593345b3.slice/crio-c351a01e8885c298bbfc34963ba846cde0a672fe03f065e0f922cd40d5948d4b WatchSource:0}: Error finding container c351a01e8885c298bbfc34963ba846cde0a672fe03f065e0f922cd40d5948d4b: Status 404 returned error can't find the container with id c351a01e8885c298bbfc34963ba846cde0a672fe03f065e0f922cd40d5948d4b Dec 05 12:30:20 crc kubenswrapper[4809]: I1205 12:30:20.880765 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8" path="/var/lib/kubelet/pods/4c1b9f1e-f3b7-41d1-bc95-29c0fc2e1ec8/volumes" Dec 05 12:30:20 crc kubenswrapper[4809]: I1205 12:30:20.900341 4809 generic.go:334] "Generic (PLEG): container finished" podID="5a0b5801-ecea-416e-a2fe-4bf8593345b3" containerID="26a062879aef1556b54f5d7f330c9356036b6ee7efb58fb9c8f8ed69aff0fcac" exitCode=0 Dec 05 12:30:20 crc kubenswrapper[4809]: I1205 12:30:20.900385 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"5a0b5801-ecea-416e-a2fe-4bf8593345b3","Type":"ContainerDied","Data":"26a062879aef1556b54f5d7f330c9356036b6ee7efb58fb9c8f8ed69aff0fcac"} Dec 05 12:30:20 crc kubenswrapper[4809]: I1205 12:30:20.900442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"5a0b5801-ecea-416e-a2fe-4bf8593345b3","Type":"ContainerStarted","Data":"c351a01e8885c298bbfc34963ba846cde0a672fe03f065e0f922cd40d5948d4b"} Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.299195 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.320038 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_5a0b5801-ecea-416e-a2fe-4bf8593345b3/mariadb-client-2/0.log" Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.350862 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.356780 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.479889 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvg7w\" (UniqueName: \"kubernetes.io/projected/5a0b5801-ecea-416e-a2fe-4bf8593345b3-kube-api-access-hvg7w\") pod \"5a0b5801-ecea-416e-a2fe-4bf8593345b3\" (UID: \"5a0b5801-ecea-416e-a2fe-4bf8593345b3\") " Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.486305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a0b5801-ecea-416e-a2fe-4bf8593345b3-kube-api-access-hvg7w" (OuterVolumeSpecName: "kube-api-access-hvg7w") pod "5a0b5801-ecea-416e-a2fe-4bf8593345b3" (UID: "5a0b5801-ecea-416e-a2fe-4bf8593345b3"). InnerVolumeSpecName "kube-api-access-hvg7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.581294 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvg7w\" (UniqueName: \"kubernetes.io/projected/5a0b5801-ecea-416e-a2fe-4bf8593345b3-kube-api-access-hvg7w\") on node \"crc\" DevicePath \"\"" Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.882969 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a0b5801-ecea-416e-a2fe-4bf8593345b3" path="/var/lib/kubelet/pods/5a0b5801-ecea-416e-a2fe-4bf8593345b3/volumes" Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.920595 4809 scope.go:117] "RemoveContainer" containerID="26a062879aef1556b54f5d7f330c9356036b6ee7efb58fb9c8f8ed69aff0fcac" Dec 05 12:30:22 crc kubenswrapper[4809]: I1205 12:30:22.920938 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 05 12:30:34 crc kubenswrapper[4809]: I1205 12:30:34.164207 4809 scope.go:117] "RemoveContainer" containerID="148e51844f6d4b3633a336f8810585ea8ac01c904ef4e6f692a9891fc72e3b75" Dec 05 12:30:34 crc kubenswrapper[4809]: I1205 12:30:34.199701 4809 scope.go:117] "RemoveContainer" containerID="7afd2813ee8499e8cf1972b63b4cbbe6b72f0d8268faf1322758dff9e62250f2" Dec 05 12:30:44 crc kubenswrapper[4809]: I1205 12:30:44.046428 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:30:44 crc kubenswrapper[4809]: I1205 12:30:44.047064 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:31:14 crc kubenswrapper[4809]: I1205 12:31:14.046292 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:31:14 crc kubenswrapper[4809]: I1205 12:31:14.047002 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.046513 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.048055 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.048223 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.049724 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.049840 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" gracePeriod=600 Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.635592 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" exitCode=0 Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.635666 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02"} Dec 05 12:31:44 crc kubenswrapper[4809]: I1205 12:31:44.635715 4809 scope.go:117] "RemoveContainer" containerID="90fcfc02adb974dcd81059abed9da13265a950cbccc4646c761fd364738d11bd" Dec 05 12:31:44 crc kubenswrapper[4809]: E1205 12:31:44.678366 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:31:45 crc kubenswrapper[4809]: I1205 12:31:45.648739 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:31:45 crc kubenswrapper[4809]: E1205 12:31:45.649615 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:32:00 crc kubenswrapper[4809]: I1205 12:32:00.872957 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:32:00 crc kubenswrapper[4809]: E1205 12:32:00.875284 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:32:14 crc kubenswrapper[4809]: I1205 12:32:14.875729 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:32:14 crc kubenswrapper[4809]: E1205 12:32:14.876507 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:32:25 crc kubenswrapper[4809]: I1205 12:32:25.872662 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:32:25 crc kubenswrapper[4809]: E1205 12:32:25.873277 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:32:39 crc kubenswrapper[4809]: I1205 12:32:39.871736 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:32:39 crc kubenswrapper[4809]: E1205 12:32:39.879127 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:32:50 crc kubenswrapper[4809]: I1205 12:32:50.872648 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:32:50 crc kubenswrapper[4809]: E1205 12:32:50.873304 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:33:04 crc kubenswrapper[4809]: I1205 12:33:04.876076 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:33:04 crc kubenswrapper[4809]: E1205 12:33:04.876934 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:33:17 crc kubenswrapper[4809]: I1205 12:33:17.873370 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:33:17 crc kubenswrapper[4809]: E1205 12:33:17.874085 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:33:31 crc kubenswrapper[4809]: I1205 12:33:31.871616 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:33:31 crc kubenswrapper[4809]: E1205 12:33:31.872329 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:33:46 crc kubenswrapper[4809]: I1205 12:33:46.872420 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:33:46 crc kubenswrapper[4809]: E1205 12:33:46.873394 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:33:57 crc kubenswrapper[4809]: I1205 12:33:57.872230 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:33:57 crc kubenswrapper[4809]: E1205 12:33:57.873055 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:34:09 crc kubenswrapper[4809]: I1205 12:34:09.871724 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:34:09 crc kubenswrapper[4809]: E1205 12:34:09.872367 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:34:24 crc kubenswrapper[4809]: I1205 12:34:24.882934 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:34:24 crc kubenswrapper[4809]: E1205 12:34:24.883707 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:34:38 crc kubenswrapper[4809]: I1205 12:34:38.871764 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:34:38 crc kubenswrapper[4809]: E1205 12:34:38.872546 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:34:52 crc kubenswrapper[4809]: I1205 12:34:52.872777 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:34:52 crc kubenswrapper[4809]: E1205 12:34:52.873518 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:35:07 crc kubenswrapper[4809]: I1205 12:35:07.872875 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:35:07 crc kubenswrapper[4809]: E1205 12:35:07.874010 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.856613 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 12:35:13 crc kubenswrapper[4809]: E1205 12:35:13.857411 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a0b5801-ecea-416e-a2fe-4bf8593345b3" containerName="mariadb-client-2" Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.857431 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a0b5801-ecea-416e-a2fe-4bf8593345b3" containerName="mariadb-client-2" Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.857614 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a0b5801-ecea-416e-a2fe-4bf8593345b3" containerName="mariadb-client-2" Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.858080 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.860003 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kdqm4" Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.897843 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.973159 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb8km\" (UniqueName: \"kubernetes.io/projected/1ea55d60-dde6-4172-9b5b-9b05bf610715-kube-api-access-bb8km\") pod \"mariadb-copy-data\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " pod="openstack/mariadb-copy-data" Dec 05 12:35:13 crc kubenswrapper[4809]: I1205 12:35:13.973284 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\") pod \"mariadb-copy-data\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " pod="openstack/mariadb-copy-data" Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.074942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\") pod \"mariadb-copy-data\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " pod="openstack/mariadb-copy-data" Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.075298 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb8km\" (UniqueName: \"kubernetes.io/projected/1ea55d60-dde6-4172-9b5b-9b05bf610715-kube-api-access-bb8km\") pod \"mariadb-copy-data\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " pod="openstack/mariadb-copy-data" Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.078053 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.078087 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\") pod \"mariadb-copy-data\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1cfe265be13e34ea0fd3891f76dce29e7722035d5d90a4c56a2d6b9efb57df3b/globalmount\"" pod="openstack/mariadb-copy-data" Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.093753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb8km\" (UniqueName: \"kubernetes.io/projected/1ea55d60-dde6-4172-9b5b-9b05bf610715-kube-api-access-bb8km\") pod \"mariadb-copy-data\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " pod="openstack/mariadb-copy-data" Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.126007 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\") pod \"mariadb-copy-data\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " pod="openstack/mariadb-copy-data" Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.178311 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 12:35:14 crc kubenswrapper[4809]: I1205 12:35:14.694950 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 12:35:15 crc kubenswrapper[4809]: I1205 12:35:15.440312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1ea55d60-dde6-4172-9b5b-9b05bf610715","Type":"ContainerStarted","Data":"36ac6df91e5181ef4f903b1a6162b47ab4ec3687b00ace0c8b1582ff7097f1f0"} Dec 05 12:35:15 crc kubenswrapper[4809]: I1205 12:35:15.440688 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1ea55d60-dde6-4172-9b5b-9b05bf610715","Type":"ContainerStarted","Data":"5bec4577a4a5d51937a4ebc32503c804c2e90100b9d66378a3657adff0cd525a"} Dec 05 12:35:15 crc kubenswrapper[4809]: I1205 12:35:15.458851 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.458830437 podStartE2EDuration="3.458830437s" podCreationTimestamp="2025-12-05 12:35:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:35:15.456454355 +0000 UTC m=+5210.847430913" watchObservedRunningTime="2025-12-05 12:35:15.458830437 +0000 UTC m=+5210.849807015" Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.205789 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.207262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.216460 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.340000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb8b9\" (UniqueName: \"kubernetes.io/projected/7f6eb497-103b-4ab6-a388-a6517ab881e0-kube-api-access-mb8b9\") pod \"mariadb-client\" (UID: \"7f6eb497-103b-4ab6-a388-a6517ab881e0\") " pod="openstack/mariadb-client" Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.441508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb8b9\" (UniqueName: \"kubernetes.io/projected/7f6eb497-103b-4ab6-a388-a6517ab881e0-kube-api-access-mb8b9\") pod \"mariadb-client\" (UID: \"7f6eb497-103b-4ab6-a388-a6517ab881e0\") " pod="openstack/mariadb-client" Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.490341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb8b9\" (UniqueName: \"kubernetes.io/projected/7f6eb497-103b-4ab6-a388-a6517ab881e0-kube-api-access-mb8b9\") pod \"mariadb-client\" (UID: \"7f6eb497-103b-4ab6-a388-a6517ab881e0\") " pod="openstack/mariadb-client" Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.531386 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:18 crc kubenswrapper[4809]: I1205 12:35:18.981743 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:19 crc kubenswrapper[4809]: W1205 12:35:19.002225 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f6eb497_103b_4ab6_a388_a6517ab881e0.slice/crio-647a9da9d871f8db337703a8fae12a1613ac6ae4281b624e4d1865807eb2b181 WatchSource:0}: Error finding container 647a9da9d871f8db337703a8fae12a1613ac6ae4281b624e4d1865807eb2b181: Status 404 returned error can't find the container with id 647a9da9d871f8db337703a8fae12a1613ac6ae4281b624e4d1865807eb2b181 Dec 05 12:35:19 crc kubenswrapper[4809]: I1205 12:35:19.478611 4809 generic.go:334] "Generic (PLEG): container finished" podID="7f6eb497-103b-4ab6-a388-a6517ab881e0" containerID="c14dc964fa4ce36371a3c919ec0a1038750bc4da5594a23069fb53748326fb7d" exitCode=0 Dec 05 12:35:19 crc kubenswrapper[4809]: I1205 12:35:19.478687 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"7f6eb497-103b-4ab6-a388-a6517ab881e0","Type":"ContainerDied","Data":"c14dc964fa4ce36371a3c919ec0a1038750bc4da5594a23069fb53748326fb7d"} Dec 05 12:35:19 crc kubenswrapper[4809]: I1205 12:35:19.478748 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"7f6eb497-103b-4ab6-a388-a6517ab881e0","Type":"ContainerStarted","Data":"647a9da9d871f8db337703a8fae12a1613ac6ae4281b624e4d1865807eb2b181"} Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.757572 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.782169 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_7f6eb497-103b-4ab6-a388-a6517ab881e0/mariadb-client/0.log" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.806368 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.813653 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.873193 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:35:20 crc kubenswrapper[4809]: E1205 12:35:20.873404 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.877003 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb8b9\" (UniqueName: \"kubernetes.io/projected/7f6eb497-103b-4ab6-a388-a6517ab881e0-kube-api-access-mb8b9\") pod \"7f6eb497-103b-4ab6-a388-a6517ab881e0\" (UID: \"7f6eb497-103b-4ab6-a388-a6517ab881e0\") " Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.882540 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f6eb497-103b-4ab6-a388-a6517ab881e0-kube-api-access-mb8b9" (OuterVolumeSpecName: "kube-api-access-mb8b9") pod "7f6eb497-103b-4ab6-a388-a6517ab881e0" (UID: "7f6eb497-103b-4ab6-a388-a6517ab881e0"). InnerVolumeSpecName "kube-api-access-mb8b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.939047 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:20 crc kubenswrapper[4809]: E1205 12:35:20.939422 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f6eb497-103b-4ab6-a388-a6517ab881e0" containerName="mariadb-client" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.939443 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f6eb497-103b-4ab6-a388-a6517ab881e0" containerName="mariadb-client" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.939605 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f6eb497-103b-4ab6-a388-a6517ab881e0" containerName="mariadb-client" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.940071 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.946668 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:20 crc kubenswrapper[4809]: I1205 12:35:20.980178 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb8b9\" (UniqueName: \"kubernetes.io/projected/7f6eb497-103b-4ab6-a388-a6517ab881e0-kube-api-access-mb8b9\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:21 crc kubenswrapper[4809]: I1205 12:35:21.081403 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xbzr\" (UniqueName: \"kubernetes.io/projected/6d0eb306-0dcc-40e7-95c4-1485954fc40d-kube-api-access-8xbzr\") pod \"mariadb-client\" (UID: \"6d0eb306-0dcc-40e7-95c4-1485954fc40d\") " pod="openstack/mariadb-client" Dec 05 12:35:21 crc kubenswrapper[4809]: I1205 12:35:21.182391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xbzr\" (UniqueName: \"kubernetes.io/projected/6d0eb306-0dcc-40e7-95c4-1485954fc40d-kube-api-access-8xbzr\") pod \"mariadb-client\" (UID: \"6d0eb306-0dcc-40e7-95c4-1485954fc40d\") " pod="openstack/mariadb-client" Dec 05 12:35:21 crc kubenswrapper[4809]: I1205 12:35:21.198422 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xbzr\" (UniqueName: \"kubernetes.io/projected/6d0eb306-0dcc-40e7-95c4-1485954fc40d-kube-api-access-8xbzr\") pod \"mariadb-client\" (UID: \"6d0eb306-0dcc-40e7-95c4-1485954fc40d\") " pod="openstack/mariadb-client" Dec 05 12:35:21 crc kubenswrapper[4809]: I1205 12:35:21.291292 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:21 crc kubenswrapper[4809]: I1205 12:35:21.494991 4809 scope.go:117] "RemoveContainer" containerID="c14dc964fa4ce36371a3c919ec0a1038750bc4da5594a23069fb53748326fb7d" Dec 05 12:35:21 crc kubenswrapper[4809]: I1205 12:35:21.495246 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:21 crc kubenswrapper[4809]: I1205 12:35:21.705944 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:21 crc kubenswrapper[4809]: W1205 12:35:21.710817 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d0eb306_0dcc_40e7_95c4_1485954fc40d.slice/crio-679d792682621bfc5ac63050da05cb801d7188d9a31abcdac66839f62515cf86 WatchSource:0}: Error finding container 679d792682621bfc5ac63050da05cb801d7188d9a31abcdac66839f62515cf86: Status 404 returned error can't find the container with id 679d792682621bfc5ac63050da05cb801d7188d9a31abcdac66839f62515cf86 Dec 05 12:35:22 crc kubenswrapper[4809]: I1205 12:35:22.503600 4809 generic.go:334] "Generic (PLEG): container finished" podID="6d0eb306-0dcc-40e7-95c4-1485954fc40d" containerID="5d0c3d20dc138d2c97c2246e5bf711d1f60dd7e71e0b8b8d4b4f63ef07534669" exitCode=0 Dec 05 12:35:22 crc kubenswrapper[4809]: I1205 12:35:22.503677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6d0eb306-0dcc-40e7-95c4-1485954fc40d","Type":"ContainerDied","Data":"5d0c3d20dc138d2c97c2246e5bf711d1f60dd7e71e0b8b8d4b4f63ef07534669"} Dec 05 12:35:22 crc kubenswrapper[4809]: I1205 12:35:22.503950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6d0eb306-0dcc-40e7-95c4-1485954fc40d","Type":"ContainerStarted","Data":"679d792682621bfc5ac63050da05cb801d7188d9a31abcdac66839f62515cf86"} Dec 05 12:35:22 crc kubenswrapper[4809]: I1205 12:35:22.885780 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f6eb497-103b-4ab6-a388-a6517ab881e0" path="/var/lib/kubelet/pods/7f6eb497-103b-4ab6-a388-a6517ab881e0/volumes" Dec 05 12:35:23 crc kubenswrapper[4809]: I1205 12:35:23.816589 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:23 crc kubenswrapper[4809]: I1205 12:35:23.837896 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_6d0eb306-0dcc-40e7-95c4-1485954fc40d/mariadb-client/0.log" Dec 05 12:35:23 crc kubenswrapper[4809]: I1205 12:35:23.863982 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:23 crc kubenswrapper[4809]: I1205 12:35:23.870815 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 05 12:35:23 crc kubenswrapper[4809]: I1205 12:35:23.932896 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xbzr\" (UniqueName: \"kubernetes.io/projected/6d0eb306-0dcc-40e7-95c4-1485954fc40d-kube-api-access-8xbzr\") pod \"6d0eb306-0dcc-40e7-95c4-1485954fc40d\" (UID: \"6d0eb306-0dcc-40e7-95c4-1485954fc40d\") " Dec 05 12:35:23 crc kubenswrapper[4809]: I1205 12:35:23.939958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d0eb306-0dcc-40e7-95c4-1485954fc40d-kube-api-access-8xbzr" (OuterVolumeSpecName: "kube-api-access-8xbzr") pod "6d0eb306-0dcc-40e7-95c4-1485954fc40d" (UID: "6d0eb306-0dcc-40e7-95c4-1485954fc40d"). InnerVolumeSpecName "kube-api-access-8xbzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:35:24 crc kubenswrapper[4809]: I1205 12:35:24.034756 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xbzr\" (UniqueName: \"kubernetes.io/projected/6d0eb306-0dcc-40e7-95c4-1485954fc40d-kube-api-access-8xbzr\") on node \"crc\" DevicePath \"\"" Dec 05 12:35:24 crc kubenswrapper[4809]: I1205 12:35:24.523786 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="679d792682621bfc5ac63050da05cb801d7188d9a31abcdac66839f62515cf86" Dec 05 12:35:24 crc kubenswrapper[4809]: I1205 12:35:24.523856 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 05 12:35:24 crc kubenswrapper[4809]: I1205 12:35:24.885230 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d0eb306-0dcc-40e7-95c4-1485954fc40d" path="/var/lib/kubelet/pods/6d0eb306-0dcc-40e7-95c4-1485954fc40d/volumes" Dec 05 12:35:35 crc kubenswrapper[4809]: I1205 12:35:35.871811 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:35:35 crc kubenswrapper[4809]: E1205 12:35:35.872577 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:35:46 crc kubenswrapper[4809]: I1205 12:35:46.872780 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:35:46 crc kubenswrapper[4809]: E1205 12:35:46.873620 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:35:57 crc kubenswrapper[4809]: I1205 12:35:57.871961 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:35:57 crc kubenswrapper[4809]: E1205 12:35:57.872589 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.249189 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 12:35:59 crc kubenswrapper[4809]: E1205 12:35:59.250346 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d0eb306-0dcc-40e7-95c4-1485954fc40d" containerName="mariadb-client" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.250370 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d0eb306-0dcc-40e7-95c4-1485954fc40d" containerName="mariadb-client" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.250981 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d0eb306-0dcc-40e7-95c4-1485954fc40d" containerName="mariadb-client" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.253361 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.257242 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-f88wv" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.260028 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.264148 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.271839 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.274602 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9353c82b-1d9c-4516-984c-855b7aaf663e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.274703 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3ffffd45-252f-409a-baf4-afe329282e61\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ffffd45-252f-409a-baf4-afe329282e61\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.274745 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353c82b-1d9c-4516-984c-855b7aaf663e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.274817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krl4l\" (UniqueName: \"kubernetes.io/projected/9353c82b-1d9c-4516-984c-855b7aaf663e-kube-api-access-krl4l\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.275034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9353c82b-1d9c-4516-984c-855b7aaf663e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.275066 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9353c82b-1d9c-4516-984c-855b7aaf663e-config\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.280368 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.281969 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.329423 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.349476 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.351112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.357216 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.376251 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9353c82b-1d9c-4516-984c-855b7aaf663e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.376325 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3ffffd45-252f-409a-baf4-afe329282e61\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ffffd45-252f-409a-baf4-afe329282e61\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.376357 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353c82b-1d9c-4516-984c-855b7aaf663e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.376400 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krl4l\" (UniqueName: \"kubernetes.io/projected/9353c82b-1d9c-4516-984c-855b7aaf663e-kube-api-access-krl4l\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.376511 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9353c82b-1d9c-4516-984c-855b7aaf663e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.376533 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9353c82b-1d9c-4516-984c-855b7aaf663e-config\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.376900 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9353c82b-1d9c-4516-984c-855b7aaf663e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.377602 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9353c82b-1d9c-4516-984c-855b7aaf663e-config\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.378079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9353c82b-1d9c-4516-984c-855b7aaf663e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.381905 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.388789 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3ffffd45-252f-409a-baf4-afe329282e61\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ffffd45-252f-409a-baf4-afe329282e61\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ff302ceb50da07c5db3e7f3357652205a574189e1c1bb32b5530de6bab71b810/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.383173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9353c82b-1d9c-4516-984c-855b7aaf663e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.393720 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krl4l\" (UniqueName: \"kubernetes.io/projected/9353c82b-1d9c-4516-984c-855b7aaf663e-kube-api-access-krl4l\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.413258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3ffffd45-252f-409a-baf4-afe329282e61\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ffffd45-252f-409a-baf4-afe329282e61\") pod \"ovsdbserver-nb-0\" (UID: \"9353c82b-1d9c-4516-984c-855b7aaf663e\") " pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.442375 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.448003 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.450400 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lgjz7" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.450679 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.454278 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.455794 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.477274 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.478790 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479426 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-config\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479595 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cb92422c-9a1b-410f-aed5-874452a39fff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb92422c-9a1b-410f-aed5-874452a39fff\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479641 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479662 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479677 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/24f0d020-eb48-422a-995d-072129b627d8-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479694 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f0d020-eb48-422a-995d-072129b627d8-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24f0d020-eb48-422a-995d-072129b627d8-config\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.479762 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.480069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwr4t\" (UniqueName: \"kubernetes.io/projected/24f0d020-eb48-422a-995d-072129b627d8-kube-api-access-zwr4t\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.480105 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7kbp\" (UniqueName: \"kubernetes.io/projected/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-kube-api-access-w7kbp\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.480129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24f0d020-eb48-422a-995d-072129b627d8-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.491481 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.492826 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.497867 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.507776 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582047 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582111 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/30ce764d-74a0-46aa-b35f-c96bdaedca59-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582141 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p67h8\" (UniqueName: \"kubernetes.io/projected/95ec104d-e045-42f0-9b19-951db84ca0cb-kube-api-access-p67h8\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-11f59930-3f70-4465-89ae-51b9130a1920\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11f59930-3f70-4465-89ae-51b9130a1920\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582208 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec104d-e045-42f0-9b19-951db84ca0cb-config\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582229 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30ce764d-74a0-46aa-b35f-c96bdaedca59-config\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwr4t\" (UniqueName: \"kubernetes.io/projected/24f0d020-eb48-422a-995d-072129b627d8-kube-api-access-zwr4t\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7kbp\" (UniqueName: \"kubernetes.io/projected/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-kube-api-access-w7kbp\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582311 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79pks\" (UniqueName: \"kubernetes.io/projected/30ce764d-74a0-46aa-b35f-c96bdaedca59-kube-api-access-79pks\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24f0d020-eb48-422a-995d-072129b627d8-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95ec104d-e045-42f0-9b19-951db84ca0cb-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ce764d-74a0-46aa-b35f-c96bdaedca59-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582387 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-config\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ce764d-74a0-46aa-b35f-c96bdaedca59-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582425 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/95ec104d-e045-42f0-9b19-951db84ca0cb-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cb92422c-9a1b-410f-aed5-874452a39fff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb92422c-9a1b-410f-aed5-874452a39fff\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582486 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/24f0d020-eb48-422a-995d-072129b627d8-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582519 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f0d020-eb48-422a-995d-072129b627d8-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582545 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ec104d-e045-42f0-9b19-951db84ca0cb-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24f0d020-eb48-422a-995d-072129b627d8-config\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.582582 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.583464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.583847 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/24f0d020-eb48-422a-995d-072129b627d8-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.584592 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-config\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.584803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24f0d020-eb48-422a-995d-072129b627d8-config\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.585750 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.586685 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.586721 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cb92422c-9a1b-410f-aed5-874452a39fff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb92422c-9a1b-410f-aed5-874452a39fff\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ccedc0ceba5e7b1935cf6bad2f90609292454da62b0b3e78e9e92c4c535e4bee/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.586978 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24f0d020-eb48-422a-995d-072129b627d8-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.587363 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.587393 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6972bef33a322e87360514d4080acdba6fb98233bf38b39077894f94709784a0/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.589796 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f0d020-eb48-422a-995d-072129b627d8-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.592194 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.600374 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwr4t\" (UniqueName: \"kubernetes.io/projected/24f0d020-eb48-422a-995d-072129b627d8-kube-api-access-zwr4t\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.601247 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7kbp\" (UniqueName: \"kubernetes.io/projected/eadb4411-40eb-47a1-994d-f73bd3d4d4b2-kube-api-access-w7kbp\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.630428 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.631045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-045f67f7-97f5-4972-aa4b-c3694b7bd9c7\") pod \"ovsdbserver-nb-2\" (UID: \"eadb4411-40eb-47a1-994d-f73bd3d4d4b2\") " pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.634626 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cb92422c-9a1b-410f-aed5-874452a39fff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb92422c-9a1b-410f-aed5-874452a39fff\") pod \"ovsdbserver-nb-1\" (UID: \"24f0d020-eb48-422a-995d-072129b627d8\") " pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.641597 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.667749 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.684022 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742477c9-f396-434f-bfd7-339df39432b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.684097 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79pks\" (UniqueName: \"kubernetes.io/projected/30ce764d-74a0-46aa-b35f-c96bdaedca59-kube-api-access-79pks\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.686246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95ec104d-e045-42f0-9b19-951db84ca0cb-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.686311 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95ec104d-e045-42f0-9b19-951db84ca0cb-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.686376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ce764d-74a0-46aa-b35f-c96bdaedca59-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687005 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/742477c9-f396-434f-bfd7-339df39432b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ce764d-74a0-46aa-b35f-c96bdaedca59-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687150 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/95ec104d-e045-42f0-9b19-951db84ca0cb-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687245 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ec104d-e045-42f0-9b19-951db84ca0cb-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742477c9-f396-434f-bfd7-339df39432b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/742477c9-f396-434f-bfd7-339df39432b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687466 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/30ce764d-74a0-46aa-b35f-c96bdaedca59-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687508 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p67h8\" (UniqueName: \"kubernetes.io/projected/95ec104d-e045-42f0-9b19-951db84ca0cb-kube-api-access-p67h8\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687619 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-11f59930-3f70-4465-89ae-51b9130a1920\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11f59930-3f70-4465-89ae-51b9130a1920\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687692 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec104d-e045-42f0-9b19-951db84ca0cb-config\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30ce764d-74a0-46aa-b35f-c96bdaedca59-config\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.687952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48vc9\" (UniqueName: \"kubernetes.io/projected/742477c9-f396-434f-bfd7-339df39432b8-kube-api-access-48vc9\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.688568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/30ce764d-74a0-46aa-b35f-c96bdaedca59-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.688574 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ce764d-74a0-46aa-b35f-c96bdaedca59-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.689609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec104d-e045-42f0-9b19-951db84ca0cb-config\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.689877 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30ce764d-74a0-46aa-b35f-c96bdaedca59-config\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.690930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ce764d-74a0-46aa-b35f-c96bdaedca59-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.692115 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.692155 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ec4bd65aa94d3733e75814f6fe3800057eac711019fd4ff8884556184fb03e59/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.693228 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.693275 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-11f59930-3f70-4465-89ae-51b9130a1920\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11f59930-3f70-4465-89ae-51b9130a1920\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/622b81398eda0eb8b940339a21f4269f9bc9a79550c61bdc51fd04d46847a96d/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.693688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ec104d-e045-42f0-9b19-951db84ca0cb-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.696197 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/95ec104d-e045-42f0-9b19-951db84ca0cb-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.707486 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79pks\" (UniqueName: \"kubernetes.io/projected/30ce764d-74a0-46aa-b35f-c96bdaedca59-kube-api-access-79pks\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.708438 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p67h8\" (UniqueName: \"kubernetes.io/projected/95ec104d-e045-42f0-9b19-951db84ca0cb-kube-api-access-p67h8\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.735772 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dafa5821-eea1-489a-aec7-f0e0a870d692\") pod \"ovsdbserver-sb-0\" (UID: \"30ce764d-74a0-46aa-b35f-c96bdaedca59\") " pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.744696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-11f59930-3f70-4465-89ae-51b9130a1920\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11f59930-3f70-4465-89ae-51b9130a1920\") pod \"ovsdbserver-sb-2\" (UID: \"95ec104d-e045-42f0-9b19-951db84ca0cb\") " pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.786689 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.789653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48vc9\" (UniqueName: \"kubernetes.io/projected/742477c9-f396-434f-bfd7-339df39432b8-kube-api-access-48vc9\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.789692 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742477c9-f396-434f-bfd7-339df39432b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.789728 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/742477c9-f396-434f-bfd7-339df39432b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.789790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742477c9-f396-434f-bfd7-339df39432b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.789810 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/742477c9-f396-434f-bfd7-339df39432b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.789831 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.790977 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/742477c9-f396-434f-bfd7-339df39432b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.793312 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742477c9-f396-434f-bfd7-339df39432b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.799340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/742477c9-f396-434f-bfd7-339df39432b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.799353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/742477c9-f396-434f-bfd7-339df39432b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.802341 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.802374 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/84112234e735ba8d207f75948f148c880612326abe3b74827becb2fd0e72d34f/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.804806 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.818936 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48vc9\" (UniqueName: \"kubernetes.io/projected/742477c9-f396-434f-bfd7-339df39432b8-kube-api-access-48vc9\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:35:59 crc kubenswrapper[4809]: I1205 12:35:59.850797 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1c7d15ca-d050-43c7-a1c0-33bf4313d237\") pod \"ovsdbserver-sb-1\" (UID: \"742477c9-f396-434f-bfd7-339df39432b8\") " pod="openstack/ovsdbserver-sb-1" Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.115382 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.231366 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 05 12:36:00 crc kubenswrapper[4809]: W1205 12:36:00.240851 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeadb4411_40eb_47a1_994d_f73bd3d4d4b2.slice/crio-feffd872330ca218944b9be8cc43084ff14d76226ea2daf701e8c3783003d119 WatchSource:0}: Error finding container feffd872330ca218944b9be8cc43084ff14d76226ea2daf701e8c3783003d119: Status 404 returned error can't find the container with id feffd872330ca218944b9be8cc43084ff14d76226ea2daf701e8c3783003d119 Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.325163 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.485368 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.656409 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 05 12:36:00 crc kubenswrapper[4809]: W1205 12:36:00.662212 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod742477c9_f396_434f_bfd7_339df39432b8.slice/crio-f4eba8f176a1af97506bbb0343ad7a4b4b0bcaf8c21f4224e016277ce65581a6 WatchSource:0}: Error finding container f4eba8f176a1af97506bbb0343ad7a4b4b0bcaf8c21f4224e016277ce65581a6: Status 404 returned error can't find the container with id f4eba8f176a1af97506bbb0343ad7a4b4b0bcaf8c21f4224e016277ce65581a6 Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.835411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"24f0d020-eb48-422a-995d-072129b627d8","Type":"ContainerStarted","Data":"263e5ba2a4b4120f58371b864729791a5f3e127d20a35c3332c37e91001e777a"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.835464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"24f0d020-eb48-422a-995d-072129b627d8","Type":"ContainerStarted","Data":"1d01b524d482a0919607b6ecea925c22836c65620bc8cbee298d04be2262508f"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.841543 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"eadb4411-40eb-47a1-994d-f73bd3d4d4b2","Type":"ContainerStarted","Data":"a6765ffe254c89776f83f652876c79023ad95057f96be8965720de3df9ad1fd0"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.841579 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"eadb4411-40eb-47a1-994d-f73bd3d4d4b2","Type":"ContainerStarted","Data":"ba7f02a38589ecfb70a2a3d7ecd44cebf8cc93ac58152e112340aa7f222ef254"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.841590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"eadb4411-40eb-47a1-994d-f73bd3d4d4b2","Type":"ContainerStarted","Data":"feffd872330ca218944b9be8cc43084ff14d76226ea2daf701e8c3783003d119"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.845323 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"95ec104d-e045-42f0-9b19-951db84ca0cb","Type":"ContainerStarted","Data":"5799499b577ffe2b559e325fc95c36238bd8653a6131e76e48d79f68d1ac3acb"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.845360 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"95ec104d-e045-42f0-9b19-951db84ca0cb","Type":"ContainerStarted","Data":"a1fc58b43aff9db8b72319f521d913e0aa190b95dbdc83c890d9d834732f1cc0"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.849644 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"742477c9-f396-434f-bfd7-339df39432b8","Type":"ContainerStarted","Data":"f4eba8f176a1af97506bbb0343ad7a4b4b0bcaf8c21f4224e016277ce65581a6"} Dec 05 12:36:00 crc kubenswrapper[4809]: I1205 12:36:00.862077 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=2.862061921 podStartE2EDuration="2.862061921s" podCreationTimestamp="2025-12-05 12:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:00.857674496 +0000 UTC m=+5256.248651054" watchObservedRunningTime="2025-12-05 12:36:00.862061921 +0000 UTC m=+5256.253038479" Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.051121 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 12:36:01 crc kubenswrapper[4809]: W1205 12:36:01.059100 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30ce764d_74a0_46aa_b35f_c96bdaedca59.slice/crio-31446b375e1669a429c98d4de7bdfde25bfe7cfa2614ab2a9b3b289460e60ae1 WatchSource:0}: Error finding container 31446b375e1669a429c98d4de7bdfde25bfe7cfa2614ab2a9b3b289460e60ae1: Status 404 returned error can't find the container with id 31446b375e1669a429c98d4de7bdfde25bfe7cfa2614ab2a9b3b289460e60ae1 Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.684228 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.863203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"95ec104d-e045-42f0-9b19-951db84ca0cb","Type":"ContainerStarted","Data":"a37e8dedb7919ce0041e2be489c087a760a81ed9e60507f0bd2baeec434e0db9"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.866318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"742477c9-f396-434f-bfd7-339df39432b8","Type":"ContainerStarted","Data":"fc2c09efcf1f53eec484264d16fc00544d6f57bc8675cf2124008dc3ecd47c12"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.866364 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"742477c9-f396-434f-bfd7-339df39432b8","Type":"ContainerStarted","Data":"ae89dc57049e1e6cedab0dc6c010d4c89c919deb10acdbcf16ffd9eb497edbdd"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.867668 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9353c82b-1d9c-4516-984c-855b7aaf663e","Type":"ContainerStarted","Data":"a4f8e48122b28a026286d50afc512dce66121be14a54ff94d603d42ae7b45d60"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.870012 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"24f0d020-eb48-422a-995d-072129b627d8","Type":"ContainerStarted","Data":"7c056710f0aea141a2dee94bd6f5dbedd4a5bc8420aafa65392996becde63d9c"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.872067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"30ce764d-74a0-46aa-b35f-c96bdaedca59","Type":"ContainerStarted","Data":"f2c092098c090db8c7ec513de6fdc816162f542e7d6f3d400616b469d60b3c7b"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.872117 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"30ce764d-74a0-46aa-b35f-c96bdaedca59","Type":"ContainerStarted","Data":"6741a34200602198a020a393439b95bcf30ba99708194d4b9396655749fc461f"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.872131 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"30ce764d-74a0-46aa-b35f-c96bdaedca59","Type":"ContainerStarted","Data":"31446b375e1669a429c98d4de7bdfde25bfe7cfa2614ab2a9b3b289460e60ae1"} Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.902019 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.901996965 podStartE2EDuration="3.901996965s" podCreationTimestamp="2025-12-05 12:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:01.8991377 +0000 UTC m=+5257.290114258" watchObservedRunningTime="2025-12-05 12:36:01.901996965 +0000 UTC m=+5257.292973523" Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.908405 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.9083827319999997 podStartE2EDuration="3.908382732s" podCreationTimestamp="2025-12-05 12:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:01.8830039 +0000 UTC m=+5257.273980458" watchObservedRunningTime="2025-12-05 12:36:01.908382732 +0000 UTC m=+5257.299359290" Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.917569 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.9175518609999997 podStartE2EDuration="3.917551861s" podCreationTimestamp="2025-12-05 12:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:01.913206997 +0000 UTC m=+5257.304183575" watchObservedRunningTime="2025-12-05 12:36:01.917551861 +0000 UTC m=+5257.308528419" Dec 05 12:36:01 crc kubenswrapper[4809]: I1205 12:36:01.935494 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.935477588 podStartE2EDuration="3.935477588s" podCreationTimestamp="2025-12-05 12:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:01.933341313 +0000 UTC m=+5257.324317871" watchObservedRunningTime="2025-12-05 12:36:01.935477588 +0000 UTC m=+5257.326454136" Dec 05 12:36:02 crc kubenswrapper[4809]: I1205 12:36:02.642576 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 05 12:36:02 crc kubenswrapper[4809]: I1205 12:36:02.668508 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 05 12:36:02 crc kubenswrapper[4809]: I1205 12:36:02.787709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 05 12:36:02 crc kubenswrapper[4809]: I1205 12:36:02.805515 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 05 12:36:02 crc kubenswrapper[4809]: I1205 12:36:02.880795 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9353c82b-1d9c-4516-984c-855b7aaf663e","Type":"ContainerStarted","Data":"94111d5295c0377f25b578c5b8109ca4890c50480489bb3eff6e93cbd0b9eb2c"} Dec 05 12:36:02 crc kubenswrapper[4809]: I1205 12:36:02.880826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9353c82b-1d9c-4516-984c-855b7aaf663e","Type":"ContainerStarted","Data":"43a31cd33ad76454f99c08b03c2fb9bc94dc8d4aeb84a35621893de2856d966f"} Dec 05 12:36:02 crc kubenswrapper[4809]: I1205 12:36:02.901343 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.901322249 podStartE2EDuration="4.901322249s" podCreationTimestamp="2025-12-05 12:35:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:02.896409381 +0000 UTC m=+5258.287385969" watchObservedRunningTime="2025-12-05 12:36:02.901322249 +0000 UTC m=+5258.292298817" Dec 05 12:36:03 crc kubenswrapper[4809]: I1205 12:36:03.117105 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 05 12:36:03 crc kubenswrapper[4809]: I1205 12:36:03.176616 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 05 12:36:03 crc kubenswrapper[4809]: I1205 12:36:03.891299 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 05 12:36:04 crc kubenswrapper[4809]: I1205 12:36:04.630957 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 05 12:36:04 crc kubenswrapper[4809]: I1205 12:36:04.642296 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 05 12:36:04 crc kubenswrapper[4809]: I1205 12:36:04.669005 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 05 12:36:04 crc kubenswrapper[4809]: I1205 12:36:04.787925 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 05 12:36:04 crc kubenswrapper[4809]: I1205 12:36:04.804945 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.153213 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.460062 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d4879d899-jdgvs"] Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.466278 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.468154 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.474028 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d4879d899-jdgvs"] Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.612521 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z98m2\" (UniqueName: \"kubernetes.io/projected/6921ac44-23bb-40a8-b0f6-20f01913542d-kube-api-access-z98m2\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.612590 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-config\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.612985 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-dns-svc\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.613052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-ovsdbserver-sb\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.631388 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.667109 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.689598 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.711281 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.714407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-ovsdbserver-sb\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.714532 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z98m2\" (UniqueName: \"kubernetes.io/projected/6921ac44-23bb-40a8-b0f6-20f01913542d-kube-api-access-z98m2\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.714586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-config\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.714698 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-dns-svc\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.715793 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-dns-svc\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.715799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-config\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.716539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-ovsdbserver-sb\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.730326 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.739367 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z98m2\" (UniqueName: \"kubernetes.io/projected/6921ac44-23bb-40a8-b0f6-20f01913542d-kube-api-access-z98m2\") pod \"dnsmasq-dns-5d4879d899-jdgvs\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.803351 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.826725 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.857263 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.960052 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 05 12:36:05 crc kubenswrapper[4809]: I1205 12:36:05.971294 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.045544 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4879d899-jdgvs"] Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.096445 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d98cd6b5c-nbnl8"] Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.098387 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.100300 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.105238 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d98cd6b5c-nbnl8"] Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.126471 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-dns-svc\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.126524 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-config\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.126572 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-sb\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.126688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmbpn\" (UniqueName: \"kubernetes.io/projected/0b7c6986-264a-4700-ae61-21b7ebb50594-kube-api-access-fmbpn\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.126747 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-nb\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.228640 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-nb\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.228749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-dns-svc\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.228778 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-config\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.228843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-sb\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.228960 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmbpn\" (UniqueName: \"kubernetes.io/projected/0b7c6986-264a-4700-ae61-21b7ebb50594-kube-api-access-fmbpn\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.229846 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-nb\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.229940 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-dns-svc\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.230256 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-config\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.230718 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-sb\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.253924 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmbpn\" (UniqueName: \"kubernetes.io/projected/0b7c6986-264a-4700-ae61-21b7ebb50594-kube-api-access-fmbpn\") pod \"dnsmasq-dns-d98cd6b5c-nbnl8\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.402017 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4879d899-jdgvs"] Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.421482 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.854745 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d98cd6b5c-nbnl8"] Dec 05 12:36:06 crc kubenswrapper[4809]: W1205 12:36:06.859186 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b7c6986_264a_4700_ae61_21b7ebb50594.slice/crio-7b6308764c1c281255ecbad4e3a21bc6a84cc9800a73b0c2d5f6f53004a335d3 WatchSource:0}: Error finding container 7b6308764c1c281255ecbad4e3a21bc6a84cc9800a73b0c2d5f6f53004a335d3: Status 404 returned error can't find the container with id 7b6308764c1c281255ecbad4e3a21bc6a84cc9800a73b0c2d5f6f53004a335d3 Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.929510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" event={"ID":"0b7c6986-264a-4700-ae61-21b7ebb50594","Type":"ContainerStarted","Data":"7b6308764c1c281255ecbad4e3a21bc6a84cc9800a73b0c2d5f6f53004a335d3"} Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.932155 4809 generic.go:334] "Generic (PLEG): container finished" podID="6921ac44-23bb-40a8-b0f6-20f01913542d" containerID="89bff43432e10e1beb8158af58a48e7f49de63f3b2b758ae448eac1eed2ba871" exitCode=0 Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.932356 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" event={"ID":"6921ac44-23bb-40a8-b0f6-20f01913542d","Type":"ContainerDied","Data":"89bff43432e10e1beb8158af58a48e7f49de63f3b2b758ae448eac1eed2ba871"} Dec 05 12:36:06 crc kubenswrapper[4809]: I1205 12:36:06.932397 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" event={"ID":"6921ac44-23bb-40a8-b0f6-20f01913542d","Type":"ContainerStarted","Data":"4c78ce4fa010d0cc33ae53a86f34b3c0f58141b37d48a5e06e5571e05a0ba8d4"} Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.007548 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.200010 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.347188 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-ovsdbserver-sb\") pod \"6921ac44-23bb-40a8-b0f6-20f01913542d\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.347262 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-config\") pod \"6921ac44-23bb-40a8-b0f6-20f01913542d\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.347304 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z98m2\" (UniqueName: \"kubernetes.io/projected/6921ac44-23bb-40a8-b0f6-20f01913542d-kube-api-access-z98m2\") pod \"6921ac44-23bb-40a8-b0f6-20f01913542d\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.347338 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-dns-svc\") pod \"6921ac44-23bb-40a8-b0f6-20f01913542d\" (UID: \"6921ac44-23bb-40a8-b0f6-20f01913542d\") " Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.351341 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6921ac44-23bb-40a8-b0f6-20f01913542d-kube-api-access-z98m2" (OuterVolumeSpecName: "kube-api-access-z98m2") pod "6921ac44-23bb-40a8-b0f6-20f01913542d" (UID: "6921ac44-23bb-40a8-b0f6-20f01913542d"). InnerVolumeSpecName "kube-api-access-z98m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.365212 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6921ac44-23bb-40a8-b0f6-20f01913542d" (UID: "6921ac44-23bb-40a8-b0f6-20f01913542d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.365950 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6921ac44-23bb-40a8-b0f6-20f01913542d" (UID: "6921ac44-23bb-40a8-b0f6-20f01913542d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.369054 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-config" (OuterVolumeSpecName: "config") pod "6921ac44-23bb-40a8-b0f6-20f01913542d" (UID: "6921ac44-23bb-40a8-b0f6-20f01913542d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.449697 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.449735 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z98m2\" (UniqueName: \"kubernetes.io/projected/6921ac44-23bb-40a8-b0f6-20f01913542d-kube-api-access-z98m2\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.449747 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.449756 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6921ac44-23bb-40a8-b0f6-20f01913542d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.945487 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.945487 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4879d899-jdgvs" event={"ID":"6921ac44-23bb-40a8-b0f6-20f01913542d","Type":"ContainerDied","Data":"4c78ce4fa010d0cc33ae53a86f34b3c0f58141b37d48a5e06e5571e05a0ba8d4"} Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.945941 4809 scope.go:117] "RemoveContainer" containerID="89bff43432e10e1beb8158af58a48e7f49de63f3b2b758ae448eac1eed2ba871" Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.947666 4809 generic.go:334] "Generic (PLEG): container finished" podID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerID="f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b" exitCode=0 Dec 05 12:36:07 crc kubenswrapper[4809]: I1205 12:36:07.948820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" event={"ID":"0b7c6986-264a-4700-ae61-21b7ebb50594","Type":"ContainerDied","Data":"f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b"} Dec 05 12:36:08 crc kubenswrapper[4809]: I1205 12:36:08.131163 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4879d899-jdgvs"] Dec 05 12:36:08 crc kubenswrapper[4809]: I1205 12:36:08.137599 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d4879d899-jdgvs"] Dec 05 12:36:08 crc kubenswrapper[4809]: I1205 12:36:08.872351 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:36:08 crc kubenswrapper[4809]: E1205 12:36:08.873652 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:36:08 crc kubenswrapper[4809]: I1205 12:36:08.882831 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6921ac44-23bb-40a8-b0f6-20f01913542d" path="/var/lib/kubelet/pods/6921ac44-23bb-40a8-b0f6-20f01913542d/volumes" Dec 05 12:36:08 crc kubenswrapper[4809]: I1205 12:36:08.955214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" event={"ID":"0b7c6986-264a-4700-ae61-21b7ebb50594","Type":"ContainerStarted","Data":"0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0"} Dec 05 12:36:08 crc kubenswrapper[4809]: I1205 12:36:08.955364 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:08 crc kubenswrapper[4809]: I1205 12:36:08.974236 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" podStartSLOduration=2.974212857 podStartE2EDuration="2.974212857s" podCreationTimestamp="2025-12-05 12:36:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:08.971295891 +0000 UTC m=+5264.362272459" watchObservedRunningTime="2025-12-05 12:36:08.974212857 +0000 UTC m=+5264.365189425" Dec 05 12:36:09 crc kubenswrapper[4809]: I1205 12:36:09.827860 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.257057 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 05 12:36:12 crc kubenswrapper[4809]: E1205 12:36:12.257723 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6921ac44-23bb-40a8-b0f6-20f01913542d" containerName="init" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.257743 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6921ac44-23bb-40a8-b0f6-20f01913542d" containerName="init" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.257999 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6921ac44-23bb-40a8-b0f6-20f01913542d" containerName="init" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.258975 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.261591 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.263896 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.428744 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvhfq\" (UniqueName: \"kubernetes.io/projected/c7fc85c2-1679-439c-ae19-9644f747c625-kube-api-access-vvhfq\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.428817 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.428852 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/c7fc85c2-1679-439c-ae19-9644f747c625-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.530727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvhfq\" (UniqueName: \"kubernetes.io/projected/c7fc85c2-1679-439c-ae19-9644f747c625-kube-api-access-vvhfq\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.531150 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.531189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/c7fc85c2-1679-439c-ae19-9644f747c625-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.536857 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.536915 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1dcddb15f8927ed057d6b793f784cad97e01cf54f89ccc7713193e513f9676f5/globalmount\"" pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.541807 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/c7fc85c2-1679-439c-ae19-9644f747c625-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.564015 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvhfq\" (UniqueName: \"kubernetes.io/projected/c7fc85c2-1679-439c-ae19-9644f747c625-kube-api-access-vvhfq\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.580258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\") pod \"ovn-copy-data\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " pod="openstack/ovn-copy-data" Dec 05 12:36:12 crc kubenswrapper[4809]: I1205 12:36:12.588802 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 12:36:13 crc kubenswrapper[4809]: I1205 12:36:13.128464 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 12:36:13 crc kubenswrapper[4809]: W1205 12:36:13.132343 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7fc85c2_1679_439c_ae19_9644f747c625.slice/crio-0da929ccf27c361136eb350442252629e10ddba1d089b918f7d3589c7a10e69e WatchSource:0}: Error finding container 0da929ccf27c361136eb350442252629e10ddba1d089b918f7d3589c7a10e69e: Status 404 returned error can't find the container with id 0da929ccf27c361136eb350442252629e10ddba1d089b918f7d3589c7a10e69e Dec 05 12:36:14 crc kubenswrapper[4809]: I1205 12:36:13.999482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"c7fc85c2-1679-439c-ae19-9644f747c625","Type":"ContainerStarted","Data":"09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4"} Dec 05 12:36:14 crc kubenswrapper[4809]: I1205 12:36:13.999890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"c7fc85c2-1679-439c-ae19-9644f747c625","Type":"ContainerStarted","Data":"0da929ccf27c361136eb350442252629e10ddba1d089b918f7d3589c7a10e69e"} Dec 05 12:36:14 crc kubenswrapper[4809]: I1205 12:36:14.016432 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.01641203 podStartE2EDuration="3.01641203s" podCreationTimestamp="2025-12-05 12:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:14.015224189 +0000 UTC m=+5269.406200787" watchObservedRunningTime="2025-12-05 12:36:14.01641203 +0000 UTC m=+5269.407388598" Dec 05 12:36:16 crc kubenswrapper[4809]: I1205 12:36:16.423396 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:16 crc kubenswrapper[4809]: I1205 12:36:16.501426 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-4zfn2"] Dec 05 12:36:16 crc kubenswrapper[4809]: I1205 12:36:16.501725 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" podUID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerName="dnsmasq-dns" containerID="cri-o://3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa" gracePeriod=10 Dec 05 12:36:16 crc kubenswrapper[4809]: I1205 12:36:16.960293 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.028544 4809 generic.go:334] "Generic (PLEG): container finished" podID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerID="3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa" exitCode=0 Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.028598 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" event={"ID":"bd51984c-a4ab-4041-8586-94c4fe70ab60","Type":"ContainerDied","Data":"3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa"} Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.028648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" event={"ID":"bd51984c-a4ab-4041-8586-94c4fe70ab60","Type":"ContainerDied","Data":"c20b9e7ec248a5dc02dd44cd0b2b5da290b877fb6ced23a50fa4376b3f3d778b"} Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.028669 4809 scope.go:117] "RemoveContainer" containerID="3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.028811 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-4zfn2" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.055983 4809 scope.go:117] "RemoveContainer" containerID="b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.084947 4809 scope.go:117] "RemoveContainer" containerID="3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa" Dec 05 12:36:17 crc kubenswrapper[4809]: E1205 12:36:17.085478 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa\": container with ID starting with 3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa not found: ID does not exist" containerID="3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.085527 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa"} err="failed to get container status \"3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa\": rpc error: code = NotFound desc = could not find container \"3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa\": container with ID starting with 3da13502b943e62c93e7fd59b2974f18ccff5862302293be48d824a7e2615dfa not found: ID does not exist" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.085559 4809 scope.go:117] "RemoveContainer" containerID="b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec" Dec 05 12:36:17 crc kubenswrapper[4809]: E1205 12:36:17.087028 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec\": container with ID starting with b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec not found: ID does not exist" containerID="b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.087071 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec"} err="failed to get container status \"b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec\": rpc error: code = NotFound desc = could not find container \"b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec\": container with ID starting with b6615d0ca3f56fd0062bc22f8c705a6a0eae264f0908a7e972e83bd554069eec not found: ID does not exist" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.110582 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-config\") pod \"bd51984c-a4ab-4041-8586-94c4fe70ab60\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.110739 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snzld\" (UniqueName: \"kubernetes.io/projected/bd51984c-a4ab-4041-8586-94c4fe70ab60-kube-api-access-snzld\") pod \"bd51984c-a4ab-4041-8586-94c4fe70ab60\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.110840 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-dns-svc\") pod \"bd51984c-a4ab-4041-8586-94c4fe70ab60\" (UID: \"bd51984c-a4ab-4041-8586-94c4fe70ab60\") " Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.115986 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd51984c-a4ab-4041-8586-94c4fe70ab60-kube-api-access-snzld" (OuterVolumeSpecName: "kube-api-access-snzld") pod "bd51984c-a4ab-4041-8586-94c4fe70ab60" (UID: "bd51984c-a4ab-4041-8586-94c4fe70ab60"). InnerVolumeSpecName "kube-api-access-snzld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.151009 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd51984c-a4ab-4041-8586-94c4fe70ab60" (UID: "bd51984c-a4ab-4041-8586-94c4fe70ab60"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.159798 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-config" (OuterVolumeSpecName: "config") pod "bd51984c-a4ab-4041-8586-94c4fe70ab60" (UID: "bd51984c-a4ab-4041-8586-94c4fe70ab60"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.213348 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snzld\" (UniqueName: \"kubernetes.io/projected/bd51984c-a4ab-4041-8586-94c4fe70ab60-kube-api-access-snzld\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.213417 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.213430 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd51984c-a4ab-4041-8586-94c4fe70ab60-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.360810 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-4zfn2"] Dec 05 12:36:17 crc kubenswrapper[4809]: I1205 12:36:17.366213 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-4zfn2"] Dec 05 12:36:17 crc kubenswrapper[4809]: E1205 12:36:17.660680 4809 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.180:50720->38.102.83.180:40539: write tcp 38.102.83.180:50720->38.102.83.180:40539: write: broken pipe Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.885769 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd51984c-a4ab-4041-8586-94c4fe70ab60" path="/var/lib/kubelet/pods/bd51984c-a4ab-4041-8586-94c4fe70ab60/volumes" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.931555 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 05 12:36:18 crc kubenswrapper[4809]: E1205 12:36:18.931950 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerName="init" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.931969 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerName="init" Dec 05 12:36:18 crc kubenswrapper[4809]: E1205 12:36:18.931981 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerName="dnsmasq-dns" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.931989 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerName="dnsmasq-dns" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.932187 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd51984c-a4ab-4041-8586-94c4fe70ab60" containerName="dnsmasq-dns" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.933224 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.935203 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.935203 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.935697 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-f7b5w" Dec 05 12:36:18 crc kubenswrapper[4809]: I1205 12:36:18.945311 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.042940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.043277 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-config\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.043341 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chtp8\" (UniqueName: \"kubernetes.io/projected/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-kube-api-access-chtp8\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.043436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.043489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-scripts\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.145133 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chtp8\" (UniqueName: \"kubernetes.io/projected/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-kube-api-access-chtp8\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.145226 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.145268 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-scripts\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.145339 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.145369 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-config\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.145803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.146290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-scripts\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.146401 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-config\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.160720 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.172593 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chtp8\" (UniqueName: \"kubernetes.io/projected/24e1ca2d-52aa-48a6-95df-c0dd39dc83fa-kube-api-access-chtp8\") pod \"ovn-northd-0\" (UID: \"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa\") " pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.277098 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 12:36:19 crc kubenswrapper[4809]: I1205 12:36:19.701381 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 12:36:19 crc kubenswrapper[4809]: W1205 12:36:19.708841 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e1ca2d_52aa_48a6_95df_c0dd39dc83fa.slice/crio-a50cda79b0267aad51bd85fd28c757ade73ba7c98410e645151b451d6364e423 WatchSource:0}: Error finding container a50cda79b0267aad51bd85fd28c757ade73ba7c98410e645151b451d6364e423: Status 404 returned error can't find the container with id a50cda79b0267aad51bd85fd28c757ade73ba7c98410e645151b451d6364e423 Dec 05 12:36:20 crc kubenswrapper[4809]: I1205 12:36:20.057285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa","Type":"ContainerStarted","Data":"e267ef8b25f189f3d333edb667d798dc439d0ce45f82abb180f2fbb051a8604d"} Dec 05 12:36:20 crc kubenswrapper[4809]: I1205 12:36:20.058462 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa","Type":"ContainerStarted","Data":"1d0093d70775234e3de54984899cc8cc6c54239c744edd9f8a1fd0591c2bbe8f"} Dec 05 12:36:20 crc kubenswrapper[4809]: I1205 12:36:20.058550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"24e1ca2d-52aa-48a6-95df-c0dd39dc83fa","Type":"ContainerStarted","Data":"a50cda79b0267aad51bd85fd28c757ade73ba7c98410e645151b451d6364e423"} Dec 05 12:36:20 crc kubenswrapper[4809]: I1205 12:36:20.058659 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 05 12:36:20 crc kubenswrapper[4809]: I1205 12:36:20.075017 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.074997014 podStartE2EDuration="2.074997014s" podCreationTimestamp="2025-12-05 12:36:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:20.07368387 +0000 UTC m=+5275.464660438" watchObservedRunningTime="2025-12-05 12:36:20.074997014 +0000 UTC m=+5275.465973572" Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.834364 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-7gk2d"] Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.835952 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.861895 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7gk2d"] Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.871760 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:36:23 crc kubenswrapper[4809]: E1205 12:36:23.872060 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.928597 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-93ae-account-create-update-hsvmz"] Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.929291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lgrc\" (UniqueName: \"kubernetes.io/projected/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-kube-api-access-6lgrc\") pod \"keystone-db-create-7gk2d\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.929351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-operator-scripts\") pod \"keystone-db-create-7gk2d\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.930157 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.934032 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 05 12:36:23 crc kubenswrapper[4809]: I1205 12:36:23.955161 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-93ae-account-create-update-hsvmz"] Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.030976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5afa0578-156c-428c-84f0-066a27a4b928-operator-scripts\") pod \"keystone-93ae-account-create-update-hsvmz\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.031055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjnwt\" (UniqueName: \"kubernetes.io/projected/5afa0578-156c-428c-84f0-066a27a4b928-kube-api-access-rjnwt\") pod \"keystone-93ae-account-create-update-hsvmz\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.031146 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lgrc\" (UniqueName: \"kubernetes.io/projected/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-kube-api-access-6lgrc\") pod \"keystone-db-create-7gk2d\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.031181 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-operator-scripts\") pod \"keystone-db-create-7gk2d\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.031895 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-operator-scripts\") pod \"keystone-db-create-7gk2d\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.051418 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lgrc\" (UniqueName: \"kubernetes.io/projected/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-kube-api-access-6lgrc\") pod \"keystone-db-create-7gk2d\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.133102 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5afa0578-156c-428c-84f0-066a27a4b928-operator-scripts\") pod \"keystone-93ae-account-create-update-hsvmz\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.133798 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjnwt\" (UniqueName: \"kubernetes.io/projected/5afa0578-156c-428c-84f0-066a27a4b928-kube-api-access-rjnwt\") pod \"keystone-93ae-account-create-update-hsvmz\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.133858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5afa0578-156c-428c-84f0-066a27a4b928-operator-scripts\") pod \"keystone-93ae-account-create-update-hsvmz\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.151467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjnwt\" (UniqueName: \"kubernetes.io/projected/5afa0578-156c-428c-84f0-066a27a4b928-kube-api-access-rjnwt\") pod \"keystone-93ae-account-create-update-hsvmz\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.164695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.267095 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.657849 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7gk2d"] Dec 05 12:36:24 crc kubenswrapper[4809]: I1205 12:36:24.743381 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-93ae-account-create-update-hsvmz"] Dec 05 12:36:24 crc kubenswrapper[4809]: W1205 12:36:24.747837 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5afa0578_156c_428c_84f0_066a27a4b928.slice/crio-0f3dea337c484983ca0555078b468159eab009d096c45cbabf864ee57a861776 WatchSource:0}: Error finding container 0f3dea337c484983ca0555078b468159eab009d096c45cbabf864ee57a861776: Status 404 returned error can't find the container with id 0f3dea337c484983ca0555078b468159eab009d096c45cbabf864ee57a861776 Dec 05 12:36:25 crc kubenswrapper[4809]: I1205 12:36:25.097273 4809 generic.go:334] "Generic (PLEG): container finished" podID="a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8" containerID="2c2828fc46bb16840eb4ac349d1b41a6d84ca0e8d8173fc8f214555df95cca8d" exitCode=0 Dec 05 12:36:25 crc kubenswrapper[4809]: I1205 12:36:25.097479 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7gk2d" event={"ID":"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8","Type":"ContainerDied","Data":"2c2828fc46bb16840eb4ac349d1b41a6d84ca0e8d8173fc8f214555df95cca8d"} Dec 05 12:36:25 crc kubenswrapper[4809]: I1205 12:36:25.097780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7gk2d" event={"ID":"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8","Type":"ContainerStarted","Data":"b4df25d44656b9c7ab559618dbdd7f93915df189a6f99165bab9fad6fb622208"} Dec 05 12:36:25 crc kubenswrapper[4809]: I1205 12:36:25.100286 4809 generic.go:334] "Generic (PLEG): container finished" podID="5afa0578-156c-428c-84f0-066a27a4b928" containerID="320a423d0e9842ec6d36250de676aa392b41e43300fc8a512d63d217ae9323ea" exitCode=0 Dec 05 12:36:25 crc kubenswrapper[4809]: I1205 12:36:25.100312 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-93ae-account-create-update-hsvmz" event={"ID":"5afa0578-156c-428c-84f0-066a27a4b928","Type":"ContainerDied","Data":"320a423d0e9842ec6d36250de676aa392b41e43300fc8a512d63d217ae9323ea"} Dec 05 12:36:25 crc kubenswrapper[4809]: I1205 12:36:25.100402 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-93ae-account-create-update-hsvmz" event={"ID":"5afa0578-156c-428c-84f0-066a27a4b928","Type":"ContainerStarted","Data":"0f3dea337c484983ca0555078b468159eab009d096c45cbabf864ee57a861776"} Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.522601 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.530516 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.676013 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lgrc\" (UniqueName: \"kubernetes.io/projected/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-kube-api-access-6lgrc\") pod \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.676066 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-operator-scripts\") pod \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\" (UID: \"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8\") " Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.676123 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjnwt\" (UniqueName: \"kubernetes.io/projected/5afa0578-156c-428c-84f0-066a27a4b928-kube-api-access-rjnwt\") pod \"5afa0578-156c-428c-84f0-066a27a4b928\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.676217 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5afa0578-156c-428c-84f0-066a27a4b928-operator-scripts\") pod \"5afa0578-156c-428c-84f0-066a27a4b928\" (UID: \"5afa0578-156c-428c-84f0-066a27a4b928\") " Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.676858 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5afa0578-156c-428c-84f0-066a27a4b928-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5afa0578-156c-428c-84f0-066a27a4b928" (UID: "5afa0578-156c-428c-84f0-066a27a4b928"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.676911 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8" (UID: "a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.677125 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5afa0578-156c-428c-84f0-066a27a4b928-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.677149 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.682502 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-kube-api-access-6lgrc" (OuterVolumeSpecName: "kube-api-access-6lgrc") pod "a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8" (UID: "a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8"). InnerVolumeSpecName "kube-api-access-6lgrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.684707 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afa0578-156c-428c-84f0-066a27a4b928-kube-api-access-rjnwt" (OuterVolumeSpecName: "kube-api-access-rjnwt") pod "5afa0578-156c-428c-84f0-066a27a4b928" (UID: "5afa0578-156c-428c-84f0-066a27a4b928"). InnerVolumeSpecName "kube-api-access-rjnwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.779527 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lgrc\" (UniqueName: \"kubernetes.io/projected/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8-kube-api-access-6lgrc\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:26 crc kubenswrapper[4809]: I1205 12:36:26.779565 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjnwt\" (UniqueName: \"kubernetes.io/projected/5afa0578-156c-428c-84f0-066a27a4b928-kube-api-access-rjnwt\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:27 crc kubenswrapper[4809]: I1205 12:36:27.121023 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7gk2d" event={"ID":"a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8","Type":"ContainerDied","Data":"b4df25d44656b9c7ab559618dbdd7f93915df189a6f99165bab9fad6fb622208"} Dec 05 12:36:27 crc kubenswrapper[4809]: I1205 12:36:27.121070 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4df25d44656b9c7ab559618dbdd7f93915df189a6f99165bab9fad6fb622208" Dec 05 12:36:27 crc kubenswrapper[4809]: I1205 12:36:27.121140 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7gk2d" Dec 05 12:36:27 crc kubenswrapper[4809]: I1205 12:36:27.123125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-93ae-account-create-update-hsvmz" event={"ID":"5afa0578-156c-428c-84f0-066a27a4b928","Type":"ContainerDied","Data":"0f3dea337c484983ca0555078b468159eab009d096c45cbabf864ee57a861776"} Dec 05 12:36:27 crc kubenswrapper[4809]: I1205 12:36:27.123453 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f3dea337c484983ca0555078b468159eab009d096c45cbabf864ee57a861776" Dec 05 12:36:27 crc kubenswrapper[4809]: I1205 12:36:27.123210 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93ae-account-create-update-hsvmz" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.346303 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.437186 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-t7kj2"] Dec 05 12:36:29 crc kubenswrapper[4809]: E1205 12:36:29.437768 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8" containerName="mariadb-database-create" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.437784 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8" containerName="mariadb-database-create" Dec 05 12:36:29 crc kubenswrapper[4809]: E1205 12:36:29.437808 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afa0578-156c-428c-84f0-066a27a4b928" containerName="mariadb-account-create-update" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.437814 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afa0578-156c-428c-84f0-066a27a4b928" containerName="mariadb-account-create-update" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.437967 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5afa0578-156c-428c-84f0-066a27a4b928" containerName="mariadb-account-create-update" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.437982 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8" containerName="mariadb-database-create" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.438472 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.439905 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.440043 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k72p7" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.440645 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.441291 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.460761 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-t7kj2"] Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.523698 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brxcw\" (UniqueName: \"kubernetes.io/projected/a9aea549-b837-4a23-bdcb-570f7a9edaab-kube-api-access-brxcw\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.523762 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-config-data\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.523860 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-combined-ca-bundle\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.625374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-combined-ca-bundle\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.625452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brxcw\" (UniqueName: \"kubernetes.io/projected/a9aea549-b837-4a23-bdcb-570f7a9edaab-kube-api-access-brxcw\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.625490 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-config-data\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.632599 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-config-data\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.633172 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-combined-ca-bundle\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.649992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brxcw\" (UniqueName: \"kubernetes.io/projected/a9aea549-b837-4a23-bdcb-570f7a9edaab-kube-api-access-brxcw\") pod \"keystone-db-sync-t7kj2\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:29 crc kubenswrapper[4809]: I1205 12:36:29.762609 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:30 crc kubenswrapper[4809]: I1205 12:36:30.217344 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-t7kj2"] Dec 05 12:36:30 crc kubenswrapper[4809]: W1205 12:36:30.222398 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9aea549_b837_4a23_bdcb_570f7a9edaab.slice/crio-9a290f83b9d21cf6b2fd65625c51b9f22cde6748be89c6d3d08fc3c17ec311fa WatchSource:0}: Error finding container 9a290f83b9d21cf6b2fd65625c51b9f22cde6748be89c6d3d08fc3c17ec311fa: Status 404 returned error can't find the container with id 9a290f83b9d21cf6b2fd65625c51b9f22cde6748be89c6d3d08fc3c17ec311fa Dec 05 12:36:31 crc kubenswrapper[4809]: I1205 12:36:31.189715 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7kj2" event={"ID":"a9aea549-b837-4a23-bdcb-570f7a9edaab","Type":"ContainerStarted","Data":"523cb049c5b8d92f1ce557d6bfb614323a3222bea8388b9e4d19d4aabd2e817b"} Dec 05 12:36:31 crc kubenswrapper[4809]: I1205 12:36:31.190010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7kj2" event={"ID":"a9aea549-b837-4a23-bdcb-570f7a9edaab","Type":"ContainerStarted","Data":"9a290f83b9d21cf6b2fd65625c51b9f22cde6748be89c6d3d08fc3c17ec311fa"} Dec 05 12:36:31 crc kubenswrapper[4809]: I1205 12:36:31.215315 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-t7kj2" podStartSLOduration=2.215288691 podStartE2EDuration="2.215288691s" podCreationTimestamp="2025-12-05 12:36:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:31.207309393 +0000 UTC m=+5286.598285971" watchObservedRunningTime="2025-12-05 12:36:31.215288691 +0000 UTC m=+5286.606265269" Dec 05 12:36:33 crc kubenswrapper[4809]: I1205 12:36:33.212202 4809 generic.go:334] "Generic (PLEG): container finished" podID="a9aea549-b837-4a23-bdcb-570f7a9edaab" containerID="523cb049c5b8d92f1ce557d6bfb614323a3222bea8388b9e4d19d4aabd2e817b" exitCode=0 Dec 05 12:36:33 crc kubenswrapper[4809]: I1205 12:36:33.212285 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7kj2" event={"ID":"a9aea549-b837-4a23-bdcb-570f7a9edaab","Type":"ContainerDied","Data":"523cb049c5b8d92f1ce557d6bfb614323a3222bea8388b9e4d19d4aabd2e817b"} Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.371591 4809 scope.go:117] "RemoveContainer" containerID="87e67d470473fa44a8d17d8f45820a07449ca12920240d23410de654f25ee331" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.398597 4809 scope.go:117] "RemoveContainer" containerID="59ea4cd6a1691e33c5431e68b39ef81cb18ea91e85643e385c3f4ca65782d1e2" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.442796 4809 scope.go:117] "RemoveContainer" containerID="19d873abe1e7ed6f397baf9fc551509b0c5960a456d6ded51994fc9326d9c3ce" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.482231 4809 scope.go:117] "RemoveContainer" containerID="ed6561da390538c84db140498382635d42afa5a4707dbda0a96a39452d28f77a" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.514023 4809 scope.go:117] "RemoveContainer" containerID="6c8088d071b24c8b405fc5952801939b6999789621074353267fd2108a946b47" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.580832 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.596958 4809 scope.go:117] "RemoveContainer" containerID="c7f93ec83840866d448cf394729f17d243c69d954bc11db1a2409da09c849a3c" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.619668 4809 scope.go:117] "RemoveContainer" containerID="bd72a84c654487405ca915cbaccafb125378031b46d40597ae4adc95e72777fb" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.728910 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-combined-ca-bundle\") pod \"a9aea549-b837-4a23-bdcb-570f7a9edaab\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.728956 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brxcw\" (UniqueName: \"kubernetes.io/projected/a9aea549-b837-4a23-bdcb-570f7a9edaab-kube-api-access-brxcw\") pod \"a9aea549-b837-4a23-bdcb-570f7a9edaab\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.729028 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-config-data\") pod \"a9aea549-b837-4a23-bdcb-570f7a9edaab\" (UID: \"a9aea549-b837-4a23-bdcb-570f7a9edaab\") " Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.736167 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9aea549-b837-4a23-bdcb-570f7a9edaab-kube-api-access-brxcw" (OuterVolumeSpecName: "kube-api-access-brxcw") pod "a9aea549-b837-4a23-bdcb-570f7a9edaab" (UID: "a9aea549-b837-4a23-bdcb-570f7a9edaab"). InnerVolumeSpecName "kube-api-access-brxcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.757522 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9aea549-b837-4a23-bdcb-570f7a9edaab" (UID: "a9aea549-b837-4a23-bdcb-570f7a9edaab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.776946 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-config-data" (OuterVolumeSpecName: "config-data") pod "a9aea549-b837-4a23-bdcb-570f7a9edaab" (UID: "a9aea549-b837-4a23-bdcb-570f7a9edaab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.830502 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.830534 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brxcw\" (UniqueName: \"kubernetes.io/projected/a9aea549-b837-4a23-bdcb-570f7a9edaab-kube-api-access-brxcw\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:34 crc kubenswrapper[4809]: I1205 12:36:34.830547 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9aea549-b837-4a23-bdcb-570f7a9edaab-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.237828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7kj2" event={"ID":"a9aea549-b837-4a23-bdcb-570f7a9edaab","Type":"ContainerDied","Data":"9a290f83b9d21cf6b2fd65625c51b9f22cde6748be89c6d3d08fc3c17ec311fa"} Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.238226 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a290f83b9d21cf6b2fd65625c51b9f22cde6748be89c6d3d08fc3c17ec311fa" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.237932 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7kj2" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.542548 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6ld69"] Dec 05 12:36:35 crc kubenswrapper[4809]: E1205 12:36:35.543069 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9aea549-b837-4a23-bdcb-570f7a9edaab" containerName="keystone-db-sync" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.543097 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9aea549-b837-4a23-bdcb-570f7a9edaab" containerName="keystone-db-sync" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.543423 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9aea549-b837-4a23-bdcb-570f7a9edaab" containerName="keystone-db-sync" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.544313 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.550337 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.550439 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k72p7" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.550980 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.551094 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.551248 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.557977 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk"] Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.559815 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.598748 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6ld69"] Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.611763 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk"] Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648093 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-config-data\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648204 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w2vb\" (UniqueName: \"kubernetes.io/projected/99018d5d-99c4-458c-b322-f650bcdeb04b-kube-api-access-5w2vb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648255 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-fernet-keys\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648368 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-dns-svc\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648401 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-combined-ca-bundle\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648448 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9qt9\" (UniqueName: \"kubernetes.io/projected/239ed119-e19d-4f3a-8254-27751bcd3426-kube-api-access-w9qt9\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648639 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-nb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648700 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-scripts\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648880 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-credential-keys\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648913 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-sb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.648949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-config\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.750878 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9qt9\" (UniqueName: \"kubernetes.io/projected/239ed119-e19d-4f3a-8254-27751bcd3426-kube-api-access-w9qt9\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.750980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-scripts\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751006 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-nb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-credential-keys\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751129 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-sb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751150 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-config\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751199 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-config-data\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w2vb\" (UniqueName: \"kubernetes.io/projected/99018d5d-99c4-458c-b322-f650bcdeb04b-kube-api-access-5w2vb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-fernet-keys\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751405 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-dns-svc\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.751430 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-combined-ca-bundle\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.752156 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-nb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.752189 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-config\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.752827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-dns-svc\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.753311 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-sb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.757566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-scripts\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.763128 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-combined-ca-bundle\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.764268 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-fernet-keys\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.770237 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-credential-keys\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.770510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-config-data\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.773336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9qt9\" (UniqueName: \"kubernetes.io/projected/239ed119-e19d-4f3a-8254-27751bcd3426-kube-api-access-w9qt9\") pod \"keystone-bootstrap-6ld69\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.773809 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w2vb\" (UniqueName: \"kubernetes.io/projected/99018d5d-99c4-458c-b322-f650bcdeb04b-kube-api-access-5w2vb\") pod \"dnsmasq-dns-6c6bcd7cbf-9dfpk\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.864522 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:35 crc kubenswrapper[4809]: I1205 12:36:35.878659 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:36 crc kubenswrapper[4809]: I1205 12:36:36.399358 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk"] Dec 05 12:36:36 crc kubenswrapper[4809]: W1205 12:36:36.404499 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99018d5d_99c4_458c_b322_f650bcdeb04b.slice/crio-3cbd6c67f2df918bd0bc1b354d9731f8763c24a3368fad007237121636d2d013 WatchSource:0}: Error finding container 3cbd6c67f2df918bd0bc1b354d9731f8763c24a3368fad007237121636d2d013: Status 404 returned error can't find the container with id 3cbd6c67f2df918bd0bc1b354d9731f8763c24a3368fad007237121636d2d013 Dec 05 12:36:36 crc kubenswrapper[4809]: I1205 12:36:36.466459 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6ld69"] Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.175824 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xzj7x"] Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.178441 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.186256 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xzj7x"] Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.269738 4809 generic.go:334] "Generic (PLEG): container finished" podID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerID="d772570f4d19ee95279fe102fb66a51c2c4823a98b5fc0669ce187179bd79477" exitCode=0 Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.269821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" event={"ID":"99018d5d-99c4-458c-b322-f650bcdeb04b","Type":"ContainerDied","Data":"d772570f4d19ee95279fe102fb66a51c2c4823a98b5fc0669ce187179bd79477"} Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.269848 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" event={"ID":"99018d5d-99c4-458c-b322-f650bcdeb04b","Type":"ContainerStarted","Data":"3cbd6c67f2df918bd0bc1b354d9731f8763c24a3368fad007237121636d2d013"} Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.273125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ld69" event={"ID":"239ed119-e19d-4f3a-8254-27751bcd3426","Type":"ContainerStarted","Data":"fb80e42d7da1e672dc0782f1c78f7440ac65eb1b2f9f612d4c2ce0d165acd6c7"} Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.273169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ld69" event={"ID":"239ed119-e19d-4f3a-8254-27751bcd3426","Type":"ContainerStarted","Data":"e981bb8461be429cc555f70f6abab1d3a0f769cb35642c016c963763a72d4778"} Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.280222 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-utilities\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.280343 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l88m\" (UniqueName: \"kubernetes.io/projected/a297b14a-1519-438d-8c81-97cd5fcc8645-kube-api-access-5l88m\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.280423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-catalog-content\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.317966 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6ld69" podStartSLOduration=2.317945475 podStartE2EDuration="2.317945475s" podCreationTimestamp="2025-12-05 12:36:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:37.31275071 +0000 UTC m=+5292.703727288" watchObservedRunningTime="2025-12-05 12:36:37.317945475 +0000 UTC m=+5292.708922033" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.381759 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-catalog-content\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.381848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-utilities\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.382302 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-catalog-content\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.382499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-utilities\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.382995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l88m\" (UniqueName: \"kubernetes.io/projected/a297b14a-1519-438d-8c81-97cd5fcc8645-kube-api-access-5l88m\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.408739 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l88m\" (UniqueName: \"kubernetes.io/projected/a297b14a-1519-438d-8c81-97cd5fcc8645-kube-api-access-5l88m\") pod \"community-operators-xzj7x\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.506075 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:37 crc kubenswrapper[4809]: I1205 12:36:37.871849 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:36:37 crc kubenswrapper[4809]: E1205 12:36:37.872850 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:36:38 crc kubenswrapper[4809]: I1205 12:36:38.109492 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xzj7x"] Dec 05 12:36:38 crc kubenswrapper[4809]: W1205 12:36:38.119190 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda297b14a_1519_438d_8c81_97cd5fcc8645.slice/crio-da3ab445fde11145f2855084afe67c9ebde8218a41a9b9e6fe403116f8b5f3af WatchSource:0}: Error finding container da3ab445fde11145f2855084afe67c9ebde8218a41a9b9e6fe403116f8b5f3af: Status 404 returned error can't find the container with id da3ab445fde11145f2855084afe67c9ebde8218a41a9b9e6fe403116f8b5f3af Dec 05 12:36:38 crc kubenswrapper[4809]: I1205 12:36:38.284386 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerStarted","Data":"46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f"} Dec 05 12:36:38 crc kubenswrapper[4809]: I1205 12:36:38.284805 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerStarted","Data":"da3ab445fde11145f2855084afe67c9ebde8218a41a9b9e6fe403116f8b5f3af"} Dec 05 12:36:38 crc kubenswrapper[4809]: I1205 12:36:38.287734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" event={"ID":"99018d5d-99c4-458c-b322-f650bcdeb04b","Type":"ContainerStarted","Data":"00390f625e401af7c7857b5734a563c4b851ff0809da3e2a56b8911a8191eaa1"} Dec 05 12:36:38 crc kubenswrapper[4809]: I1205 12:36:38.335612 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" podStartSLOduration=3.335587877 podStartE2EDuration="3.335587877s" podCreationTimestamp="2025-12-05 12:36:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:38.328918173 +0000 UTC m=+5293.719894751" watchObservedRunningTime="2025-12-05 12:36:38.335587877 +0000 UTC m=+5293.726564435" Dec 05 12:36:39 crc kubenswrapper[4809]: I1205 12:36:39.294794 4809 generic.go:334] "Generic (PLEG): container finished" podID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerID="46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f" exitCode=0 Dec 05 12:36:39 crc kubenswrapper[4809]: I1205 12:36:39.294890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerDied","Data":"46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f"} Dec 05 12:36:39 crc kubenswrapper[4809]: I1205 12:36:39.295344 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:39 crc kubenswrapper[4809]: I1205 12:36:39.297513 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:36:40 crc kubenswrapper[4809]: I1205 12:36:40.305612 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerStarted","Data":"7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748"} Dec 05 12:36:40 crc kubenswrapper[4809]: I1205 12:36:40.307156 4809 generic.go:334] "Generic (PLEG): container finished" podID="239ed119-e19d-4f3a-8254-27751bcd3426" containerID="fb80e42d7da1e672dc0782f1c78f7440ac65eb1b2f9f612d4c2ce0d165acd6c7" exitCode=0 Dec 05 12:36:40 crc kubenswrapper[4809]: I1205 12:36:40.307273 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ld69" event={"ID":"239ed119-e19d-4f3a-8254-27751bcd3426","Type":"ContainerDied","Data":"fb80e42d7da1e672dc0782f1c78f7440ac65eb1b2f9f612d4c2ce0d165acd6c7"} Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.331696 4809 generic.go:334] "Generic (PLEG): container finished" podID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerID="7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748" exitCode=0 Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.331832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerDied","Data":"7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748"} Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.676023 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.770595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-config-data\") pod \"239ed119-e19d-4f3a-8254-27751bcd3426\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.770664 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-fernet-keys\") pod \"239ed119-e19d-4f3a-8254-27751bcd3426\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.770763 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-credential-keys\") pod \"239ed119-e19d-4f3a-8254-27751bcd3426\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.770833 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-scripts\") pod \"239ed119-e19d-4f3a-8254-27751bcd3426\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.770867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-combined-ca-bundle\") pod \"239ed119-e19d-4f3a-8254-27751bcd3426\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.770926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9qt9\" (UniqueName: \"kubernetes.io/projected/239ed119-e19d-4f3a-8254-27751bcd3426-kube-api-access-w9qt9\") pod \"239ed119-e19d-4f3a-8254-27751bcd3426\" (UID: \"239ed119-e19d-4f3a-8254-27751bcd3426\") " Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.777040 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "239ed119-e19d-4f3a-8254-27751bcd3426" (UID: "239ed119-e19d-4f3a-8254-27751bcd3426"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.777082 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239ed119-e19d-4f3a-8254-27751bcd3426-kube-api-access-w9qt9" (OuterVolumeSpecName: "kube-api-access-w9qt9") pod "239ed119-e19d-4f3a-8254-27751bcd3426" (UID: "239ed119-e19d-4f3a-8254-27751bcd3426"). InnerVolumeSpecName "kube-api-access-w9qt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.777673 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-scripts" (OuterVolumeSpecName: "scripts") pod "239ed119-e19d-4f3a-8254-27751bcd3426" (UID: "239ed119-e19d-4f3a-8254-27751bcd3426"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.778333 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "239ed119-e19d-4f3a-8254-27751bcd3426" (UID: "239ed119-e19d-4f3a-8254-27751bcd3426"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.798937 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-config-data" (OuterVolumeSpecName: "config-data") pod "239ed119-e19d-4f3a-8254-27751bcd3426" (UID: "239ed119-e19d-4f3a-8254-27751bcd3426"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.801961 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "239ed119-e19d-4f3a-8254-27751bcd3426" (UID: "239ed119-e19d-4f3a-8254-27751bcd3426"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.873422 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.873447 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.873456 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.873481 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9qt9\" (UniqueName: \"kubernetes.io/projected/239ed119-e19d-4f3a-8254-27751bcd3426-kube-api-access-w9qt9\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.873490 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:41 crc kubenswrapper[4809]: I1205 12:36:41.873498 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/239ed119-e19d-4f3a-8254-27751bcd3426-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.341021 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerStarted","Data":"a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945"} Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.342463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ld69" event={"ID":"239ed119-e19d-4f3a-8254-27751bcd3426","Type":"ContainerDied","Data":"e981bb8461be429cc555f70f6abab1d3a0f769cb35642c016c963763a72d4778"} Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.342487 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e981bb8461be429cc555f70f6abab1d3a0f769cb35642c016c963763a72d4778" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.342521 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ld69" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.368332 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xzj7x" podStartSLOduration=2.931890544 podStartE2EDuration="5.368301662s" podCreationTimestamp="2025-12-05 12:36:37 +0000 UTC" firstStartedPulling="2025-12-05 12:36:39.297053715 +0000 UTC m=+5294.688030273" lastFinishedPulling="2025-12-05 12:36:41.733464833 +0000 UTC m=+5297.124441391" observedRunningTime="2025-12-05 12:36:42.358521476 +0000 UTC m=+5297.749498044" watchObservedRunningTime="2025-12-05 12:36:42.368301662 +0000 UTC m=+5297.759278260" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.423970 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6ld69"] Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.432903 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6ld69"] Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.527885 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bhwls"] Dec 05 12:36:42 crc kubenswrapper[4809]: E1205 12:36:42.528338 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239ed119-e19d-4f3a-8254-27751bcd3426" containerName="keystone-bootstrap" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.528364 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="239ed119-e19d-4f3a-8254-27751bcd3426" containerName="keystone-bootstrap" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.528567 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="239ed119-e19d-4f3a-8254-27751bcd3426" containerName="keystone-bootstrap" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.529342 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.531739 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.531971 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.532015 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.532599 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k72p7" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.538066 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bhwls"] Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.541986 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.584308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-config-data\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.584355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z444l\" (UniqueName: \"kubernetes.io/projected/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-kube-api-access-z444l\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.584397 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-combined-ca-bundle\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.584703 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-scripts\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.584979 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-fernet-keys\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.585074 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-credential-keys\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.687713 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-config-data\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.687765 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z444l\" (UniqueName: \"kubernetes.io/projected/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-kube-api-access-z444l\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.687827 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-combined-ca-bundle\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.687886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-scripts\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.687955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-fernet-keys\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.687996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-credential-keys\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.692586 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-credential-keys\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.692746 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-fernet-keys\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.693557 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-scripts\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.694072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-combined-ca-bundle\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.697823 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-config-data\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.709340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z444l\" (UniqueName: \"kubernetes.io/projected/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-kube-api-access-z444l\") pod \"keystone-bootstrap-bhwls\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.882814 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239ed119-e19d-4f3a-8254-27751bcd3426" path="/var/lib/kubelet/pods/239ed119-e19d-4f3a-8254-27751bcd3426/volumes" Dec 05 12:36:42 crc kubenswrapper[4809]: I1205 12:36:42.903666 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:43 crc kubenswrapper[4809]: I1205 12:36:43.347263 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bhwls"] Dec 05 12:36:44 crc kubenswrapper[4809]: I1205 12:36:44.362208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bhwls" event={"ID":"dea6d817-5ff8-4e2f-b39d-b61187f79ec8","Type":"ContainerStarted","Data":"6c8ae1eb2adf92e436e69ee7ebdb5cac204463a961f97e0b1a9a838bfff47152"} Dec 05 12:36:44 crc kubenswrapper[4809]: I1205 12:36:44.362537 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bhwls" event={"ID":"dea6d817-5ff8-4e2f-b39d-b61187f79ec8","Type":"ContainerStarted","Data":"1baccb8a0506c345be14a2dfb1b6473f526a31d9fdb86d40c8e8b3596e4808a5"} Dec 05 12:36:44 crc kubenswrapper[4809]: I1205 12:36:44.382058 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bhwls" podStartSLOduration=2.382034426 podStartE2EDuration="2.382034426s" podCreationTimestamp="2025-12-05 12:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:44.376284266 +0000 UTC m=+5299.767260844" watchObservedRunningTime="2025-12-05 12:36:44.382034426 +0000 UTC m=+5299.773010984" Dec 05 12:36:45 crc kubenswrapper[4809]: I1205 12:36:45.880816 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:36:45 crc kubenswrapper[4809]: I1205 12:36:45.957958 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d98cd6b5c-nbnl8"] Dec 05 12:36:45 crc kubenswrapper[4809]: I1205 12:36:45.958238 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerName="dnsmasq-dns" containerID="cri-o://0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0" gracePeriod=10 Dec 05 12:36:46 crc kubenswrapper[4809]: I1205 12:36:46.422137 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.12:5353: connect: connection refused" Dec 05 12:36:46 crc kubenswrapper[4809]: I1205 12:36:46.927723 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.067469 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmbpn\" (UniqueName: \"kubernetes.io/projected/0b7c6986-264a-4700-ae61-21b7ebb50594-kube-api-access-fmbpn\") pod \"0b7c6986-264a-4700-ae61-21b7ebb50594\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.067561 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-dns-svc\") pod \"0b7c6986-264a-4700-ae61-21b7ebb50594\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.067610 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-sb\") pod \"0b7c6986-264a-4700-ae61-21b7ebb50594\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.067680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-config\") pod \"0b7c6986-264a-4700-ae61-21b7ebb50594\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.067731 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-nb\") pod \"0b7c6986-264a-4700-ae61-21b7ebb50594\" (UID: \"0b7c6986-264a-4700-ae61-21b7ebb50594\") " Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.073761 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7c6986-264a-4700-ae61-21b7ebb50594-kube-api-access-fmbpn" (OuterVolumeSpecName: "kube-api-access-fmbpn") pod "0b7c6986-264a-4700-ae61-21b7ebb50594" (UID: "0b7c6986-264a-4700-ae61-21b7ebb50594"). InnerVolumeSpecName "kube-api-access-fmbpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.108793 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0b7c6986-264a-4700-ae61-21b7ebb50594" (UID: "0b7c6986-264a-4700-ae61-21b7ebb50594"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.115392 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b7c6986-264a-4700-ae61-21b7ebb50594" (UID: "0b7c6986-264a-4700-ae61-21b7ebb50594"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.116177 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b7c6986-264a-4700-ae61-21b7ebb50594" (UID: "0b7c6986-264a-4700-ae61-21b7ebb50594"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.126362 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-config" (OuterVolumeSpecName: "config") pod "0b7c6986-264a-4700-ae61-21b7ebb50594" (UID: "0b7c6986-264a-4700-ae61-21b7ebb50594"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.170137 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.170180 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.170195 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmbpn\" (UniqueName: \"kubernetes.io/projected/0b7c6986-264a-4700-ae61-21b7ebb50594-kube-api-access-fmbpn\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.170209 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.170223 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b7c6986-264a-4700-ae61-21b7ebb50594-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.389171 4809 generic.go:334] "Generic (PLEG): container finished" podID="dea6d817-5ff8-4e2f-b39d-b61187f79ec8" containerID="6c8ae1eb2adf92e436e69ee7ebdb5cac204463a961f97e0b1a9a838bfff47152" exitCode=0 Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.389277 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bhwls" event={"ID":"dea6d817-5ff8-4e2f-b39d-b61187f79ec8","Type":"ContainerDied","Data":"6c8ae1eb2adf92e436e69ee7ebdb5cac204463a961f97e0b1a9a838bfff47152"} Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.391705 4809 generic.go:334] "Generic (PLEG): container finished" podID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerID="0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0" exitCode=0 Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.391756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" event={"ID":"0b7c6986-264a-4700-ae61-21b7ebb50594","Type":"ContainerDied","Data":"0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0"} Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.391785 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" event={"ID":"0b7c6986-264a-4700-ae61-21b7ebb50594","Type":"ContainerDied","Data":"7b6308764c1c281255ecbad4e3a21bc6a84cc9800a73b0c2d5f6f53004a335d3"} Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.391803 4809 scope.go:117] "RemoveContainer" containerID="0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.394920 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d98cd6b5c-nbnl8" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.415873 4809 scope.go:117] "RemoveContainer" containerID="f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.435338 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d98cd6b5c-nbnl8"] Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.438408 4809 scope.go:117] "RemoveContainer" containerID="0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0" Dec 05 12:36:47 crc kubenswrapper[4809]: E1205 12:36:47.438875 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0\": container with ID starting with 0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0 not found: ID does not exist" containerID="0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.438922 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0"} err="failed to get container status \"0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0\": rpc error: code = NotFound desc = could not find container \"0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0\": container with ID starting with 0da3c8fef58d79f675901c640808bc9305be1509d774d3dcc0b804a74395cdb0 not found: ID does not exist" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.438950 4809 scope.go:117] "RemoveContainer" containerID="f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b" Dec 05 12:36:47 crc kubenswrapper[4809]: E1205 12:36:47.439473 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b\": container with ID starting with f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b not found: ID does not exist" containerID="f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.439573 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b"} err="failed to get container status \"f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b\": rpc error: code = NotFound desc = could not find container \"f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b\": container with ID starting with f6f11b6b4ff08266a438ce06f549731fbd695fe982e6286ddb6c206df28fa91b not found: ID does not exist" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.444083 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d98cd6b5c-nbnl8"] Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.506298 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.506594 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:47 crc kubenswrapper[4809]: I1205 12:36:47.549412 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.465564 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.523920 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xzj7x"] Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.713960 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.872537 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.883098 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" path="/var/lib/kubelet/pods/0b7c6986-264a-4700-ae61-21b7ebb50594/volumes" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.896903 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z444l\" (UniqueName: \"kubernetes.io/projected/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-kube-api-access-z444l\") pod \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.897004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-combined-ca-bundle\") pod \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.897035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-config-data\") pod \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.897082 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-credential-keys\") pod \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.897106 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-fernet-keys\") pod \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.897140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-scripts\") pod \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\" (UID: \"dea6d817-5ff8-4e2f-b39d-b61187f79ec8\") " Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.902684 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "dea6d817-5ff8-4e2f-b39d-b61187f79ec8" (UID: "dea6d817-5ff8-4e2f-b39d-b61187f79ec8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.902761 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-kube-api-access-z444l" (OuterVolumeSpecName: "kube-api-access-z444l") pod "dea6d817-5ff8-4e2f-b39d-b61187f79ec8" (UID: "dea6d817-5ff8-4e2f-b39d-b61187f79ec8"). InnerVolumeSpecName "kube-api-access-z444l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.902756 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-scripts" (OuterVolumeSpecName: "scripts") pod "dea6d817-5ff8-4e2f-b39d-b61187f79ec8" (UID: "dea6d817-5ff8-4e2f-b39d-b61187f79ec8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.904190 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "dea6d817-5ff8-4e2f-b39d-b61187f79ec8" (UID: "dea6d817-5ff8-4e2f-b39d-b61187f79ec8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.924360 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dea6d817-5ff8-4e2f-b39d-b61187f79ec8" (UID: "dea6d817-5ff8-4e2f-b39d-b61187f79ec8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:48 crc kubenswrapper[4809]: I1205 12:36:48.924452 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-config-data" (OuterVolumeSpecName: "config-data") pod "dea6d817-5ff8-4e2f-b39d-b61187f79ec8" (UID: "dea6d817-5ff8-4e2f-b39d-b61187f79ec8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.000832 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.000885 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.000899 4809 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.000909 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.000920 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.000933 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z444l\" (UniqueName: \"kubernetes.io/projected/dea6d817-5ff8-4e2f-b39d-b61187f79ec8-kube-api-access-z444l\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.410746 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"0d29fa2887bb281af96ca44548b0644914998505ad459688152d496f4d53f7af"} Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.412833 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bhwls" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.412914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bhwls" event={"ID":"dea6d817-5ff8-4e2f-b39d-b61187f79ec8","Type":"ContainerDied","Data":"1baccb8a0506c345be14a2dfb1b6473f526a31d9fdb86d40c8e8b3596e4808a5"} Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.412989 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1baccb8a0506c345be14a2dfb1b6473f526a31d9fdb86d40c8e8b3596e4808a5" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.527448 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c849bd98c-xwxgc"] Dec 05 12:36:49 crc kubenswrapper[4809]: E1205 12:36:49.527870 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerName="init" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.527888 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerName="init" Dec 05 12:36:49 crc kubenswrapper[4809]: E1205 12:36:49.527917 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerName="dnsmasq-dns" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.527927 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerName="dnsmasq-dns" Dec 05 12:36:49 crc kubenswrapper[4809]: E1205 12:36:49.527965 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea6d817-5ff8-4e2f-b39d-b61187f79ec8" containerName="keystone-bootstrap" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.527973 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea6d817-5ff8-4e2f-b39d-b61187f79ec8" containerName="keystone-bootstrap" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.528152 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea6d817-5ff8-4e2f-b39d-b61187f79ec8" containerName="keystone-bootstrap" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.528168 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7c6986-264a-4700-ae61-21b7ebb50594" containerName="dnsmasq-dns" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.528904 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.531357 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k72p7" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.531606 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.531777 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.531897 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.542133 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c849bd98c-xwxgc"] Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.611096 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-scripts\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.611174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-fernet-keys\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.611209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkwk8\" (UniqueName: \"kubernetes.io/projected/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-kube-api-access-zkwk8\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.611271 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-combined-ca-bundle\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.611289 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-credential-keys\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.611308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-config-data\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.712801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkwk8\" (UniqueName: \"kubernetes.io/projected/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-kube-api-access-zkwk8\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.712898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-combined-ca-bundle\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.712916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-credential-keys\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.712936 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-config-data\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.712958 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-scripts\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.713654 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-fernet-keys\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.718087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-credential-keys\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.718161 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-fernet-keys\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.718267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-scripts\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.719173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-combined-ca-bundle\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.736686 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-config-data\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.737433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkwk8\" (UniqueName: \"kubernetes.io/projected/7039dc11-ca87-46f6-ad62-503f8c4a4bbb-kube-api-access-zkwk8\") pod \"keystone-6c849bd98c-xwxgc\" (UID: \"7039dc11-ca87-46f6-ad62-503f8c4a4bbb\") " pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:49 crc kubenswrapper[4809]: I1205 12:36:49.864617 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:50 crc kubenswrapper[4809]: I1205 12:36:50.331082 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c849bd98c-xwxgc"] Dec 05 12:36:50 crc kubenswrapper[4809]: W1205 12:36:50.332801 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7039dc11_ca87_46f6_ad62_503f8c4a4bbb.slice/crio-58cc521597ee9d7ee64dd26695806a347c4d8c5a4a73a6f5a6789f82e1870f79 WatchSource:0}: Error finding container 58cc521597ee9d7ee64dd26695806a347c4d8c5a4a73a6f5a6789f82e1870f79: Status 404 returned error can't find the container with id 58cc521597ee9d7ee64dd26695806a347c4d8c5a4a73a6f5a6789f82e1870f79 Dec 05 12:36:50 crc kubenswrapper[4809]: I1205 12:36:50.424881 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xzj7x" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="registry-server" containerID="cri-o://a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945" gracePeriod=2 Dec 05 12:36:50 crc kubenswrapper[4809]: I1205 12:36:50.425371 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c849bd98c-xwxgc" event={"ID":"7039dc11-ca87-46f6-ad62-503f8c4a4bbb","Type":"ContainerStarted","Data":"58cc521597ee9d7ee64dd26695806a347c4d8c5a4a73a6f5a6789f82e1870f79"} Dec 05 12:36:50 crc kubenswrapper[4809]: I1205 12:36:50.849220 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.033995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l88m\" (UniqueName: \"kubernetes.io/projected/a297b14a-1519-438d-8c81-97cd5fcc8645-kube-api-access-5l88m\") pod \"a297b14a-1519-438d-8c81-97cd5fcc8645\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.034500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-utilities\") pod \"a297b14a-1519-438d-8c81-97cd5fcc8645\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.034556 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-catalog-content\") pod \"a297b14a-1519-438d-8c81-97cd5fcc8645\" (UID: \"a297b14a-1519-438d-8c81-97cd5fcc8645\") " Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.036014 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-utilities" (OuterVolumeSpecName: "utilities") pod "a297b14a-1519-438d-8c81-97cd5fcc8645" (UID: "a297b14a-1519-438d-8c81-97cd5fcc8645"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.046983 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a297b14a-1519-438d-8c81-97cd5fcc8645-kube-api-access-5l88m" (OuterVolumeSpecName: "kube-api-access-5l88m") pod "a297b14a-1519-438d-8c81-97cd5fcc8645" (UID: "a297b14a-1519-438d-8c81-97cd5fcc8645"). InnerVolumeSpecName "kube-api-access-5l88m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.090845 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a297b14a-1519-438d-8c81-97cd5fcc8645" (UID: "a297b14a-1519-438d-8c81-97cd5fcc8645"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.136327 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.136365 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l88m\" (UniqueName: \"kubernetes.io/projected/a297b14a-1519-438d-8c81-97cd5fcc8645-kube-api-access-5l88m\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.136378 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a297b14a-1519-438d-8c81-97cd5fcc8645-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.440469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c849bd98c-xwxgc" event={"ID":"7039dc11-ca87-46f6-ad62-503f8c4a4bbb","Type":"ContainerStarted","Data":"c05ec6069e6fa7d1701d82d2074bdb15142fbcacc3e477bb7d4e1bd53b8b997a"} Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.440846 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.446926 4809 generic.go:334] "Generic (PLEG): container finished" podID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerID="a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945" exitCode=0 Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.447017 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerDied","Data":"a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945"} Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.447102 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzj7x" event={"ID":"a297b14a-1519-438d-8c81-97cd5fcc8645","Type":"ContainerDied","Data":"da3ab445fde11145f2855084afe67c9ebde8218a41a9b9e6fe403116f8b5f3af"} Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.447131 4809 scope.go:117] "RemoveContainer" containerID="a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.447607 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzj7x" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.469931 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c849bd98c-xwxgc" podStartSLOduration=2.469908656 podStartE2EDuration="2.469908656s" podCreationTimestamp="2025-12-05 12:36:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:36:51.469476585 +0000 UTC m=+5306.860453173" watchObservedRunningTime="2025-12-05 12:36:51.469908656 +0000 UTC m=+5306.860885234" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.481968 4809 scope.go:117] "RemoveContainer" containerID="7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.506506 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xzj7x"] Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.516035 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xzj7x"] Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.529543 4809 scope.go:117] "RemoveContainer" containerID="46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.548946 4809 scope.go:117] "RemoveContainer" containerID="a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945" Dec 05 12:36:51 crc kubenswrapper[4809]: E1205 12:36:51.549509 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945\": container with ID starting with a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945 not found: ID does not exist" containerID="a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.549548 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945"} err="failed to get container status \"a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945\": rpc error: code = NotFound desc = could not find container \"a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945\": container with ID starting with a7469669d707a1cd0d22c0530e995090a0ce28becfe02c8993baf0424bbff945 not found: ID does not exist" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.549682 4809 scope.go:117] "RemoveContainer" containerID="7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748" Dec 05 12:36:51 crc kubenswrapper[4809]: E1205 12:36:51.549907 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748\": container with ID starting with 7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748 not found: ID does not exist" containerID="7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.550059 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748"} err="failed to get container status \"7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748\": rpc error: code = NotFound desc = could not find container \"7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748\": container with ID starting with 7ca5eaac0c7f695dc595eae7319a2c26cfc4d7a583dc1d063a607b302750d748 not found: ID does not exist" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.550214 4809 scope.go:117] "RemoveContainer" containerID="46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f" Dec 05 12:36:51 crc kubenswrapper[4809]: E1205 12:36:51.550488 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f\": container with ID starting with 46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f not found: ID does not exist" containerID="46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f" Dec 05 12:36:51 crc kubenswrapper[4809]: I1205 12:36:51.550523 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f"} err="failed to get container status \"46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f\": rpc error: code = NotFound desc = could not find container \"46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f\": container with ID starting with 46d7fe8247ac3e3df5000273811db789d724300e549d751540fa4e1f0d5b059f not found: ID does not exist" Dec 05 12:36:52 crc kubenswrapper[4809]: I1205 12:36:52.883372 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" path="/var/lib/kubelet/pods/a297b14a-1519-438d-8c81-97cd5fcc8645/volumes" Dec 05 12:37:21 crc kubenswrapper[4809]: I1205 12:37:21.548420 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c849bd98c-xwxgc" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.835183 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 12:37:23 crc kubenswrapper[4809]: E1205 12:37:23.835879 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="registry-server" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.835896 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="registry-server" Dec 05 12:37:23 crc kubenswrapper[4809]: E1205 12:37:23.835910 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="extract-utilities" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.835917 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="extract-utilities" Dec 05 12:37:23 crc kubenswrapper[4809]: E1205 12:37:23.835936 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="extract-content" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.835943 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="extract-content" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.836145 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a297b14a-1519-438d-8c81-97cd5fcc8645" containerName="registry-server" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.836831 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.839096 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-h2sks" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.839464 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.840737 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.848950 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqgcb\" (UniqueName: \"kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.849034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config-secret\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.849069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.854436 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.872669 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.881214 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 12:37:23 crc kubenswrapper[4809]: E1205 12:37:23.886336 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-hqgcb openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="d3279a06-c606-4a3a-856f-5af14b18fac0" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.907589 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.908655 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.917292 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.950923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5992h\" (UniqueName: \"kubernetes.io/projected/73acb568-95c4-4d30-81bd-1056b887ea65-kube-api-access-5992h\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.951006 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.951031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config-secret\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.951053 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config-secret\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.951090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.951185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqgcb\" (UniqueName: \"kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: E1205 12:37:23.953372 4809 projected.go:194] Error preparing data for projected volume kube-api-access-hqgcb for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (d3279a06-c606-4a3a-856f-5af14b18fac0) does not match the UID in record. The object might have been deleted and then recreated Dec 05 12:37:23 crc kubenswrapper[4809]: E1205 12:37:23.953435 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb podName:d3279a06-c606-4a3a-856f-5af14b18fac0 nodeName:}" failed. No retries permitted until 2025-12-05 12:37:24.453414579 +0000 UTC m=+5339.844391137 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hqgcb" (UniqueName: "kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb") pod "openstackclient" (UID: "d3279a06-c606-4a3a-856f-5af14b18fac0") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (d3279a06-c606-4a3a-856f-5af14b18fac0) does not match the UID in record. The object might have been deleted and then recreated Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.954539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:23 crc kubenswrapper[4809]: I1205 12:37:23.961857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config-secret\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.053521 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5992h\" (UniqueName: \"kubernetes.io/projected/73acb568-95c4-4d30-81bd-1056b887ea65-kube-api-access-5992h\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.053586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.053612 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config-secret\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.054697 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.059168 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config-secret\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.070474 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5992h\" (UniqueName: \"kubernetes.io/projected/73acb568-95c4-4d30-81bd-1056b887ea65-kube-api-access-5992h\") pod \"openstackclient\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.229063 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.460973 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqgcb\" (UniqueName: \"kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb\") pod \"openstackclient\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: E1205 12:37:24.463175 4809 projected.go:194] Error preparing data for projected volume kube-api-access-hqgcb for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (d3279a06-c606-4a3a-856f-5af14b18fac0) does not match the UID in record. The object might have been deleted and then recreated Dec 05 12:37:24 crc kubenswrapper[4809]: E1205 12:37:24.463243 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb podName:d3279a06-c606-4a3a-856f-5af14b18fac0 nodeName:}" failed. No retries permitted until 2025-12-05 12:37:25.463226106 +0000 UTC m=+5340.854202664 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-hqgcb" (UniqueName: "kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb") pod "openstackclient" (UID: "d3279a06-c606-4a3a-856f-5af14b18fac0") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (d3279a06-c606-4a3a-856f-5af14b18fac0) does not match the UID in record. The object might have been deleted and then recreated Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.717008 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.735884 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.735884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"73acb568-95c4-4d30-81bd-1056b887ea65","Type":"ContainerStarted","Data":"5d7527eba188730e5594975d7114eca097d544d1f4dcbf2b38b7cb4460756e25"} Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.739191 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d3279a06-c606-4a3a-856f-5af14b18fac0" podUID="73acb568-95c4-4d30-81bd-1056b887ea65" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.749415 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.766322 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqgcb\" (UniqueName: \"kubernetes.io/projected/d3279a06-c606-4a3a-856f-5af14b18fac0-kube-api-access-hqgcb\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.867759 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config\") pod \"d3279a06-c606-4a3a-856f-5af14b18fac0\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.867988 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config-secret\") pod \"d3279a06-c606-4a3a-856f-5af14b18fac0\" (UID: \"d3279a06-c606-4a3a-856f-5af14b18fac0\") " Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.868475 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d3279a06-c606-4a3a-856f-5af14b18fac0" (UID: "d3279a06-c606-4a3a-856f-5af14b18fac0"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.871746 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d3279a06-c606-4a3a-856f-5af14b18fac0" (UID: "d3279a06-c606-4a3a-856f-5af14b18fac0"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.887689 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3279a06-c606-4a3a-856f-5af14b18fac0" path="/var/lib/kubelet/pods/d3279a06-c606-4a3a-856f-5af14b18fac0/volumes" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.969396 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:24 crc kubenswrapper[4809]: I1205 12:37:24.969433 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d3279a06-c606-4a3a-856f-5af14b18fac0-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:37:25 crc kubenswrapper[4809]: I1205 12:37:25.749362 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:37:25 crc kubenswrapper[4809]: I1205 12:37:25.749359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"73acb568-95c4-4d30-81bd-1056b887ea65","Type":"ContainerStarted","Data":"9c6e66048455719a05e49fd6068d11524a5db88f7e8bbc27f4952110fcfee101"} Dec 05 12:37:25 crc kubenswrapper[4809]: I1205 12:37:25.777028 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.777009413 podStartE2EDuration="2.777009413s" podCreationTimestamp="2025-12-05 12:37:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:37:25.776422428 +0000 UTC m=+5341.167399016" watchObservedRunningTime="2025-12-05 12:37:25.777009413 +0000 UTC m=+5341.167985981" Dec 05 12:37:25 crc kubenswrapper[4809]: I1205 12:37:25.779819 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d3279a06-c606-4a3a-856f-5af14b18fac0" podUID="73acb568-95c4-4d30-81bd-1056b887ea65" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.349798 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rjkrf"] Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.379258 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.386970 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rjkrf"] Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.428315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-utilities\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.428462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-catalog-content\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.428589 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p88n\" (UniqueName: \"kubernetes.io/projected/7257cb65-e5de-425a-b6b1-c739c566d063-kube-api-access-7p88n\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.529197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-catalog-content\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.529274 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p88n\" (UniqueName: \"kubernetes.io/projected/7257cb65-e5de-425a-b6b1-c739c566d063-kube-api-access-7p88n\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.529324 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-utilities\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.529956 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-utilities\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.529950 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-catalog-content\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.557046 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p88n\" (UniqueName: \"kubernetes.io/projected/7257cb65-e5de-425a-b6b1-c739c566d063-kube-api-access-7p88n\") pod \"certified-operators-rjkrf\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:15 crc kubenswrapper[4809]: I1205 12:38:15.710838 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:16 crc kubenswrapper[4809]: I1205 12:38:16.004031 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rjkrf"] Dec 05 12:38:16 crc kubenswrapper[4809]: I1205 12:38:16.165350 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjkrf" event={"ID":"7257cb65-e5de-425a-b6b1-c739c566d063","Type":"ContainerStarted","Data":"e13db119a6e8b8da99442dc046a9de1547bdfd38eb76f982cbae78f2d579144b"} Dec 05 12:38:17 crc kubenswrapper[4809]: I1205 12:38:17.176357 4809 generic.go:334] "Generic (PLEG): container finished" podID="7257cb65-e5de-425a-b6b1-c739c566d063" containerID="16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789" exitCode=0 Dec 05 12:38:17 crc kubenswrapper[4809]: I1205 12:38:17.176544 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjkrf" event={"ID":"7257cb65-e5de-425a-b6b1-c739c566d063","Type":"ContainerDied","Data":"16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789"} Dec 05 12:38:19 crc kubenswrapper[4809]: I1205 12:38:19.193577 4809 generic.go:334] "Generic (PLEG): container finished" podID="7257cb65-e5de-425a-b6b1-c739c566d063" containerID="7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f" exitCode=0 Dec 05 12:38:19 crc kubenswrapper[4809]: I1205 12:38:19.193890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjkrf" event={"ID":"7257cb65-e5de-425a-b6b1-c739c566d063","Type":"ContainerDied","Data":"7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f"} Dec 05 12:38:20 crc kubenswrapper[4809]: I1205 12:38:20.205977 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjkrf" event={"ID":"7257cb65-e5de-425a-b6b1-c739c566d063","Type":"ContainerStarted","Data":"6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98"} Dec 05 12:38:20 crc kubenswrapper[4809]: I1205 12:38:20.227042 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rjkrf" podStartSLOduration=2.722186579 podStartE2EDuration="5.227017308s" podCreationTimestamp="2025-12-05 12:38:15 +0000 UTC" firstStartedPulling="2025-12-05 12:38:17.178332361 +0000 UTC m=+5392.569308919" lastFinishedPulling="2025-12-05 12:38:19.68316308 +0000 UTC m=+5395.074139648" observedRunningTime="2025-12-05 12:38:20.226127274 +0000 UTC m=+5395.617103852" watchObservedRunningTime="2025-12-05 12:38:20.227017308 +0000 UTC m=+5395.617993886" Dec 05 12:38:25 crc kubenswrapper[4809]: I1205 12:38:25.711504 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:25 crc kubenswrapper[4809]: I1205 12:38:25.712267 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:25 crc kubenswrapper[4809]: I1205 12:38:25.793283 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:26 crc kubenswrapper[4809]: I1205 12:38:26.292513 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:26 crc kubenswrapper[4809]: I1205 12:38:26.341503 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rjkrf"] Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.261233 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rjkrf" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="registry-server" containerID="cri-o://6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98" gracePeriod=2 Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.667232 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.758873 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-catalog-content\") pod \"7257cb65-e5de-425a-b6b1-c739c566d063\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.759614 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p88n\" (UniqueName: \"kubernetes.io/projected/7257cb65-e5de-425a-b6b1-c739c566d063-kube-api-access-7p88n\") pod \"7257cb65-e5de-425a-b6b1-c739c566d063\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.760683 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-utilities\") pod \"7257cb65-e5de-425a-b6b1-c739c566d063\" (UID: \"7257cb65-e5de-425a-b6b1-c739c566d063\") " Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.761494 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-utilities" (OuterVolumeSpecName: "utilities") pod "7257cb65-e5de-425a-b6b1-c739c566d063" (UID: "7257cb65-e5de-425a-b6b1-c739c566d063"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.761905 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.765748 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7257cb65-e5de-425a-b6b1-c739c566d063-kube-api-access-7p88n" (OuterVolumeSpecName: "kube-api-access-7p88n") pod "7257cb65-e5de-425a-b6b1-c739c566d063" (UID: "7257cb65-e5de-425a-b6b1-c739c566d063"). InnerVolumeSpecName "kube-api-access-7p88n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.863464 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p88n\" (UniqueName: \"kubernetes.io/projected/7257cb65-e5de-425a-b6b1-c739c566d063-kube-api-access-7p88n\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.897788 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7257cb65-e5de-425a-b6b1-c739c566d063" (UID: "7257cb65-e5de-425a-b6b1-c739c566d063"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:38:28 crc kubenswrapper[4809]: I1205 12:38:28.964755 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7257cb65-e5de-425a-b6b1-c739c566d063-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.270465 4809 generic.go:334] "Generic (PLEG): container finished" podID="7257cb65-e5de-425a-b6b1-c739c566d063" containerID="6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98" exitCode=0 Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.270505 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjkrf" event={"ID":"7257cb65-e5de-425a-b6b1-c739c566d063","Type":"ContainerDied","Data":"6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98"} Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.270511 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rjkrf" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.270540 4809 scope.go:117] "RemoveContainer" containerID="6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.270530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rjkrf" event={"ID":"7257cb65-e5de-425a-b6b1-c739c566d063","Type":"ContainerDied","Data":"e13db119a6e8b8da99442dc046a9de1547bdfd38eb76f982cbae78f2d579144b"} Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.297952 4809 scope.go:117] "RemoveContainer" containerID="7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.317716 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rjkrf"] Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.327805 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rjkrf"] Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.345763 4809 scope.go:117] "RemoveContainer" containerID="16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.363736 4809 scope.go:117] "RemoveContainer" containerID="6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98" Dec 05 12:38:29 crc kubenswrapper[4809]: E1205 12:38:29.364210 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98\": container with ID starting with 6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98 not found: ID does not exist" containerID="6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.364274 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98"} err="failed to get container status \"6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98\": rpc error: code = NotFound desc = could not find container \"6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98\": container with ID starting with 6f00581f5b43a95f27a7a0b375703e6f9643c97c7d699d4243fd2e7d80726f98 not found: ID does not exist" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.364310 4809 scope.go:117] "RemoveContainer" containerID="7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f" Dec 05 12:38:29 crc kubenswrapper[4809]: E1205 12:38:29.364677 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f\": container with ID starting with 7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f not found: ID does not exist" containerID="7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.364743 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f"} err="failed to get container status \"7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f\": rpc error: code = NotFound desc = could not find container \"7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f\": container with ID starting with 7362f0781f996f9c559052aa0fde06fbc7de04706aca9666e45161d617e31c3f not found: ID does not exist" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.364789 4809 scope.go:117] "RemoveContainer" containerID="16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789" Dec 05 12:38:29 crc kubenswrapper[4809]: E1205 12:38:29.365137 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789\": container with ID starting with 16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789 not found: ID does not exist" containerID="16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789" Dec 05 12:38:29 crc kubenswrapper[4809]: I1205 12:38:29.365170 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789"} err="failed to get container status \"16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789\": rpc error: code = NotFound desc = could not find container \"16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789\": container with ID starting with 16b87c6664e173659c7e9da5f6bf2d150c09bb724aab891296768602d269a789 not found: ID does not exist" Dec 05 12:38:30 crc kubenswrapper[4809]: I1205 12:38:30.889436 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" path="/var/lib/kubelet/pods/7257cb65-e5de-425a-b6b1-c739c566d063/volumes" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.379667 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sn7zk"] Dec 05 12:38:57 crc kubenswrapper[4809]: E1205 12:38:57.380741 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="extract-content" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.380762 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="extract-content" Dec 05 12:38:57 crc kubenswrapper[4809]: E1205 12:38:57.380780 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="registry-server" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.380787 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="registry-server" Dec 05 12:38:57 crc kubenswrapper[4809]: E1205 12:38:57.380813 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="extract-utilities" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.380820 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="extract-utilities" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.380989 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7257cb65-e5de-425a-b6b1-c739c566d063" containerName="registry-server" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.382230 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.391805 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sn7zk"] Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.484716 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bt52\" (UniqueName: \"kubernetes.io/projected/be2dd974-bccc-4d5f-9450-780cb4e91b0d-kube-api-access-5bt52\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.484846 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-utilities\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.484909 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-catalog-content\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.586475 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bt52\" (UniqueName: \"kubernetes.io/projected/be2dd974-bccc-4d5f-9450-780cb4e91b0d-kube-api-access-5bt52\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.586867 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-utilities\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.586919 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-catalog-content\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.587355 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-utilities\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.587385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-catalog-content\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.607194 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bt52\" (UniqueName: \"kubernetes.io/projected/be2dd974-bccc-4d5f-9450-780cb4e91b0d-kube-api-access-5bt52\") pod \"redhat-operators-sn7zk\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:57 crc kubenswrapper[4809]: I1205 12:38:57.710904 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:38:58 crc kubenswrapper[4809]: I1205 12:38:58.195623 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sn7zk"] Dec 05 12:38:58 crc kubenswrapper[4809]: I1205 12:38:58.528245 4809 generic.go:334] "Generic (PLEG): container finished" podID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerID="5089cae48dfdd2513ce14536089cf8bfb4921bf78f1ee72639f94a7e3173504a" exitCode=0 Dec 05 12:38:58 crc kubenswrapper[4809]: I1205 12:38:58.528583 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sn7zk" event={"ID":"be2dd974-bccc-4d5f-9450-780cb4e91b0d","Type":"ContainerDied","Data":"5089cae48dfdd2513ce14536089cf8bfb4921bf78f1ee72639f94a7e3173504a"} Dec 05 12:38:58 crc kubenswrapper[4809]: I1205 12:38:58.528616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sn7zk" event={"ID":"be2dd974-bccc-4d5f-9450-780cb4e91b0d","Type":"ContainerStarted","Data":"d61b051ce2a2b78f08892192810c774687872e6a4e8b9e12b5ea9ad4965d24b8"} Dec 05 12:39:00 crc kubenswrapper[4809]: I1205 12:39:00.545856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sn7zk" event={"ID":"be2dd974-bccc-4d5f-9450-780cb4e91b0d","Type":"ContainerStarted","Data":"55834004d1931f88ba4ff456fd01b4d00d7af0d82bef62f34150f955105f29aa"} Dec 05 12:39:01 crc kubenswrapper[4809]: I1205 12:39:01.555039 4809 generic.go:334] "Generic (PLEG): container finished" podID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerID="55834004d1931f88ba4ff456fd01b4d00d7af0d82bef62f34150f955105f29aa" exitCode=0 Dec 05 12:39:01 crc kubenswrapper[4809]: I1205 12:39:01.555092 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sn7zk" event={"ID":"be2dd974-bccc-4d5f-9450-780cb4e91b0d","Type":"ContainerDied","Data":"55834004d1931f88ba4ff456fd01b4d00d7af0d82bef62f34150f955105f29aa"} Dec 05 12:39:03 crc kubenswrapper[4809]: I1205 12:39:03.574761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sn7zk" event={"ID":"be2dd974-bccc-4d5f-9450-780cb4e91b0d","Type":"ContainerStarted","Data":"fa2dc1ce8d93d2bd4aae915e0174dde1d0bd907bb94cf673b800fa843a298220"} Dec 05 12:39:03 crc kubenswrapper[4809]: I1205 12:39:03.599278 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sn7zk" podStartSLOduration=2.541420615 podStartE2EDuration="6.599257245s" podCreationTimestamp="2025-12-05 12:38:57 +0000 UTC" firstStartedPulling="2025-12-05 12:38:58.530126265 +0000 UTC m=+5433.921102823" lastFinishedPulling="2025-12-05 12:39:02.587962895 +0000 UTC m=+5437.978939453" observedRunningTime="2025-12-05 12:39:03.598246238 +0000 UTC m=+5438.989222796" watchObservedRunningTime="2025-12-05 12:39:03.599257245 +0000 UTC m=+5438.990233803" Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.711275 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.711612 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.826710 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rl8qz"] Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.833294 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.837576 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rl8qz"] Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.925908 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-cba5-account-create-update-kmm8r"] Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.927060 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.930203 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 05 12:39:07 crc kubenswrapper[4809]: I1205 12:39:07.942389 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-cba5-account-create-update-kmm8r"] Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.013542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfsng\" (UniqueName: \"kubernetes.io/projected/ea306765-6369-43b9-85ba-cb8046c8c489-kube-api-access-dfsng\") pod \"barbican-cba5-account-create-update-kmm8r\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.013899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec410a7-ff97-4658-a11f-dcf2de86d676-operator-scripts\") pod \"barbican-db-create-rl8qz\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.014028 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dgmm\" (UniqueName: \"kubernetes.io/projected/7ec410a7-ff97-4658-a11f-dcf2de86d676-kube-api-access-8dgmm\") pod \"barbican-db-create-rl8qz\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.014191 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea306765-6369-43b9-85ba-cb8046c8c489-operator-scripts\") pod \"barbican-cba5-account-create-update-kmm8r\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.115607 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea306765-6369-43b9-85ba-cb8046c8c489-operator-scripts\") pod \"barbican-cba5-account-create-update-kmm8r\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.115720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfsng\" (UniqueName: \"kubernetes.io/projected/ea306765-6369-43b9-85ba-cb8046c8c489-kube-api-access-dfsng\") pod \"barbican-cba5-account-create-update-kmm8r\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.115755 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec410a7-ff97-4658-a11f-dcf2de86d676-operator-scripts\") pod \"barbican-db-create-rl8qz\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.115819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dgmm\" (UniqueName: \"kubernetes.io/projected/7ec410a7-ff97-4658-a11f-dcf2de86d676-kube-api-access-8dgmm\") pod \"barbican-db-create-rl8qz\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.116524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea306765-6369-43b9-85ba-cb8046c8c489-operator-scripts\") pod \"barbican-cba5-account-create-update-kmm8r\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.116843 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec410a7-ff97-4658-a11f-dcf2de86d676-operator-scripts\") pod \"barbican-db-create-rl8qz\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.139662 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dgmm\" (UniqueName: \"kubernetes.io/projected/7ec410a7-ff97-4658-a11f-dcf2de86d676-kube-api-access-8dgmm\") pod \"barbican-db-create-rl8qz\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.140212 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfsng\" (UniqueName: \"kubernetes.io/projected/ea306765-6369-43b9-85ba-cb8046c8c489-kube-api-access-dfsng\") pod \"barbican-cba5-account-create-update-kmm8r\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.153674 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.249680 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.652691 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rl8qz"] Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.779844 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sn7zk" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="registry-server" probeResult="failure" output=< Dec 05 12:39:08 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 12:39:08 crc kubenswrapper[4809]: > Dec 05 12:39:08 crc kubenswrapper[4809]: I1205 12:39:08.791585 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-cba5-account-create-update-kmm8r"] Dec 05 12:39:09 crc kubenswrapper[4809]: I1205 12:39:09.636617 4809 generic.go:334] "Generic (PLEG): container finished" podID="ea306765-6369-43b9-85ba-cb8046c8c489" containerID="83b504189f41cf2291b81c492f373c8072043fb0be2b9079538fe4f8c64ab7a7" exitCode=0 Dec 05 12:39:09 crc kubenswrapper[4809]: I1205 12:39:09.636703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cba5-account-create-update-kmm8r" event={"ID":"ea306765-6369-43b9-85ba-cb8046c8c489","Type":"ContainerDied","Data":"83b504189f41cf2291b81c492f373c8072043fb0be2b9079538fe4f8c64ab7a7"} Dec 05 12:39:09 crc kubenswrapper[4809]: I1205 12:39:09.636969 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cba5-account-create-update-kmm8r" event={"ID":"ea306765-6369-43b9-85ba-cb8046c8c489","Type":"ContainerStarted","Data":"19e524e4567bae6547202ed5d16d6299d4fc41fbb21edda09d37b76c50925687"} Dec 05 12:39:09 crc kubenswrapper[4809]: I1205 12:39:09.638580 4809 generic.go:334] "Generic (PLEG): container finished" podID="7ec410a7-ff97-4658-a11f-dcf2de86d676" containerID="4ccb7977ea8f4338e5eb061e90aad234bc5a6744cd141eb8c3be3b6c57df63bd" exitCode=0 Dec 05 12:39:09 crc kubenswrapper[4809]: I1205 12:39:09.638719 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rl8qz" event={"ID":"7ec410a7-ff97-4658-a11f-dcf2de86d676","Type":"ContainerDied","Data":"4ccb7977ea8f4338e5eb061e90aad234bc5a6744cd141eb8c3be3b6c57df63bd"} Dec 05 12:39:09 crc kubenswrapper[4809]: I1205 12:39:09.638820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rl8qz" event={"ID":"7ec410a7-ff97-4658-a11f-dcf2de86d676","Type":"ContainerStarted","Data":"d5a8aeb3708ba04494baa1fda730a1d92be19d279bf4e1bb145aa35d6cc48d10"} Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.015370 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.026832 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.168481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec410a7-ff97-4658-a11f-dcf2de86d676-operator-scripts\") pod \"7ec410a7-ff97-4658-a11f-dcf2de86d676\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.168539 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfsng\" (UniqueName: \"kubernetes.io/projected/ea306765-6369-43b9-85ba-cb8046c8c489-kube-api-access-dfsng\") pod \"ea306765-6369-43b9-85ba-cb8046c8c489\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.168601 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dgmm\" (UniqueName: \"kubernetes.io/projected/7ec410a7-ff97-4658-a11f-dcf2de86d676-kube-api-access-8dgmm\") pod \"7ec410a7-ff97-4658-a11f-dcf2de86d676\" (UID: \"7ec410a7-ff97-4658-a11f-dcf2de86d676\") " Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.168736 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea306765-6369-43b9-85ba-cb8046c8c489-operator-scripts\") pod \"ea306765-6369-43b9-85ba-cb8046c8c489\" (UID: \"ea306765-6369-43b9-85ba-cb8046c8c489\") " Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.169947 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec410a7-ff97-4658-a11f-dcf2de86d676-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ec410a7-ff97-4658-a11f-dcf2de86d676" (UID: "7ec410a7-ff97-4658-a11f-dcf2de86d676"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.170691 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea306765-6369-43b9-85ba-cb8046c8c489-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea306765-6369-43b9-85ba-cb8046c8c489" (UID: "ea306765-6369-43b9-85ba-cb8046c8c489"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.171078 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea306765-6369-43b9-85ba-cb8046c8c489-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.171139 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec410a7-ff97-4658-a11f-dcf2de86d676-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.175305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ec410a7-ff97-4658-a11f-dcf2de86d676-kube-api-access-8dgmm" (OuterVolumeSpecName: "kube-api-access-8dgmm") pod "7ec410a7-ff97-4658-a11f-dcf2de86d676" (UID: "7ec410a7-ff97-4658-a11f-dcf2de86d676"). InnerVolumeSpecName "kube-api-access-8dgmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.175390 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea306765-6369-43b9-85ba-cb8046c8c489-kube-api-access-dfsng" (OuterVolumeSpecName: "kube-api-access-dfsng") pod "ea306765-6369-43b9-85ba-cb8046c8c489" (UID: "ea306765-6369-43b9-85ba-cb8046c8c489"). InnerVolumeSpecName "kube-api-access-dfsng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.272932 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfsng\" (UniqueName: \"kubernetes.io/projected/ea306765-6369-43b9-85ba-cb8046c8c489-kube-api-access-dfsng\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.273247 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dgmm\" (UniqueName: \"kubernetes.io/projected/7ec410a7-ff97-4658-a11f-dcf2de86d676-kube-api-access-8dgmm\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.672706 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rl8qz" event={"ID":"7ec410a7-ff97-4658-a11f-dcf2de86d676","Type":"ContainerDied","Data":"d5a8aeb3708ba04494baa1fda730a1d92be19d279bf4e1bb145aa35d6cc48d10"} Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.672751 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5a8aeb3708ba04494baa1fda730a1d92be19d279bf4e1bb145aa35d6cc48d10" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.672825 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rl8qz" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.683895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cba5-account-create-update-kmm8r" event={"ID":"ea306765-6369-43b9-85ba-cb8046c8c489","Type":"ContainerDied","Data":"19e524e4567bae6547202ed5d16d6299d4fc41fbb21edda09d37b76c50925687"} Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.683958 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19e524e4567bae6547202ed5d16d6299d4fc41fbb21edda09d37b76c50925687" Dec 05 12:39:11 crc kubenswrapper[4809]: I1205 12:39:11.684044 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cba5-account-create-update-kmm8r" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.319300 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-hz87m"] Dec 05 12:39:13 crc kubenswrapper[4809]: E1205 12:39:13.324546 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea306765-6369-43b9-85ba-cb8046c8c489" containerName="mariadb-account-create-update" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.324584 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea306765-6369-43b9-85ba-cb8046c8c489" containerName="mariadb-account-create-update" Dec 05 12:39:13 crc kubenswrapper[4809]: E1205 12:39:13.324622 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec410a7-ff97-4658-a11f-dcf2de86d676" containerName="mariadb-database-create" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.324644 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec410a7-ff97-4658-a11f-dcf2de86d676" containerName="mariadb-database-create" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.325254 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ec410a7-ff97-4658-a11f-dcf2de86d676" containerName="mariadb-database-create" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.325291 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea306765-6369-43b9-85ba-cb8046c8c489" containerName="mariadb-account-create-update" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.326456 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.333223 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.335760 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-w4npk" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.355771 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hz87m"] Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.406779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-combined-ca-bundle\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.406912 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-db-sync-config-data\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.508169 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwzvw\" (UniqueName: \"kubernetes.io/projected/d65186f3-26d3-4c29-8bc8-41d0b57891a1-kube-api-access-qwzvw\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.508507 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-combined-ca-bundle\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.508712 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-db-sync-config-data\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.514104 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-combined-ca-bundle\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.514502 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-db-sync-config-data\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.610494 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwzvw\" (UniqueName: \"kubernetes.io/projected/d65186f3-26d3-4c29-8bc8-41d0b57891a1-kube-api-access-qwzvw\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.634045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwzvw\" (UniqueName: \"kubernetes.io/projected/d65186f3-26d3-4c29-8bc8-41d0b57891a1-kube-api-access-qwzvw\") pod \"barbican-db-sync-hz87m\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.656027 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:13 crc kubenswrapper[4809]: I1205 12:39:13.920953 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hz87m"] Dec 05 12:39:14 crc kubenswrapper[4809]: I1205 12:39:14.047046 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:39:14 crc kubenswrapper[4809]: I1205 12:39:14.047698 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:39:14 crc kubenswrapper[4809]: I1205 12:39:14.711134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hz87m" event={"ID":"d65186f3-26d3-4c29-8bc8-41d0b57891a1","Type":"ContainerStarted","Data":"34f363b0a314f287ba07633977d15c53b76537029fb877f6cb4d720bf1040750"} Dec 05 12:39:15 crc kubenswrapper[4809]: I1205 12:39:15.722442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hz87m" event={"ID":"d65186f3-26d3-4c29-8bc8-41d0b57891a1","Type":"ContainerStarted","Data":"2e74766c5d5b2b280f7d796bfbe66eb20ff3d9a4c1350f688cebedd4f4956608"} Dec 05 12:39:15 crc kubenswrapper[4809]: I1205 12:39:15.748436 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-hz87m" podStartSLOduration=2.748407082 podStartE2EDuration="2.748407082s" podCreationTimestamp="2025-12-05 12:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:39:15.744657722 +0000 UTC m=+5451.135634280" watchObservedRunningTime="2025-12-05 12:39:15.748407082 +0000 UTC m=+5451.139383640" Dec 05 12:39:16 crc kubenswrapper[4809]: I1205 12:39:16.732981 4809 generic.go:334] "Generic (PLEG): container finished" podID="d65186f3-26d3-4c29-8bc8-41d0b57891a1" containerID="2e74766c5d5b2b280f7d796bfbe66eb20ff3d9a4c1350f688cebedd4f4956608" exitCode=0 Dec 05 12:39:16 crc kubenswrapper[4809]: I1205 12:39:16.733036 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hz87m" event={"ID":"d65186f3-26d3-4c29-8bc8-41d0b57891a1","Type":"ContainerDied","Data":"2e74766c5d5b2b280f7d796bfbe66eb20ff3d9a4c1350f688cebedd4f4956608"} Dec 05 12:39:17 crc kubenswrapper[4809]: I1205 12:39:17.757171 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:39:17 crc kubenswrapper[4809]: I1205 12:39:17.805063 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.005176 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sn7zk"] Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.113463 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.199470 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-combined-ca-bundle\") pod \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.199537 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwzvw\" (UniqueName: \"kubernetes.io/projected/d65186f3-26d3-4c29-8bc8-41d0b57891a1-kube-api-access-qwzvw\") pod \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.199566 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-db-sync-config-data\") pod \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\" (UID: \"d65186f3-26d3-4c29-8bc8-41d0b57891a1\") " Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.205573 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65186f3-26d3-4c29-8bc8-41d0b57891a1-kube-api-access-qwzvw" (OuterVolumeSpecName: "kube-api-access-qwzvw") pod "d65186f3-26d3-4c29-8bc8-41d0b57891a1" (UID: "d65186f3-26d3-4c29-8bc8-41d0b57891a1"). InnerVolumeSpecName "kube-api-access-qwzvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.206057 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d65186f3-26d3-4c29-8bc8-41d0b57891a1" (UID: "d65186f3-26d3-4c29-8bc8-41d0b57891a1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.224064 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d65186f3-26d3-4c29-8bc8-41d0b57891a1" (UID: "d65186f3-26d3-4c29-8bc8-41d0b57891a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.304155 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.304206 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwzvw\" (UniqueName: \"kubernetes.io/projected/d65186f3-26d3-4c29-8bc8-41d0b57891a1-kube-api-access-qwzvw\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.304219 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d65186f3-26d3-4c29-8bc8-41d0b57891a1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.752891 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hz87m" Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.752881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hz87m" event={"ID":"d65186f3-26d3-4c29-8bc8-41d0b57891a1","Type":"ContainerDied","Data":"34f363b0a314f287ba07633977d15c53b76537029fb877f6cb4d720bf1040750"} Dec 05 12:39:18 crc kubenswrapper[4809]: I1205 12:39:18.753039 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34f363b0a314f287ba07633977d15c53b76537029fb877f6cb4d720bf1040750" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.388089 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5b9d486759-h8xx5"] Dec 05 12:39:19 crc kubenswrapper[4809]: E1205 12:39:19.388892 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d65186f3-26d3-4c29-8bc8-41d0b57891a1" containerName="barbican-db-sync" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.388915 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d65186f3-26d3-4c29-8bc8-41d0b57891a1" containerName="barbican-db-sync" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.389155 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d65186f3-26d3-4c29-8bc8-41d0b57891a1" containerName="barbican-db-sync" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.396576 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.399870 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.400095 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.405521 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-w4npk" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.414847 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b9d486759-h8xx5"] Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.424149 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-config-data-custom\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.424195 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-logs\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.424264 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-combined-ca-bundle\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.424348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6t74\" (UniqueName: \"kubernetes.io/projected/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-kube-api-access-m6t74\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.424385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-config-data\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.463831 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-78f54888c6-ht5kn"] Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.465224 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.470260 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.498964 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-78f54888c6-ht5kn"] Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.515697 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-697b7cf9cc-qwr99"] Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.517314 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.525876 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-nb\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.525919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.525952 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-config-data-custom\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.525971 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-logs\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.525995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-combined-ca-bundle\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526031 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-config-data-custom\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-combined-ca-bundle\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526091 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6t74\" (UniqueName: \"kubernetes.io/projected/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-kube-api-access-m6t74\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526119 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-config-data\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526139 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6726bf-d6e9-43fa-828f-c78798598723-logs\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526251 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-dns-svc\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526557 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-logs\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.526897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-sb\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.527052 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnx2q\" (UniqueName: \"kubernetes.io/projected/4fbf0c52-301f-4d29-a333-e8dc4a25448c-kube-api-access-bnx2q\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.527131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-config-data\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.527165 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fh4c\" (UniqueName: \"kubernetes.io/projected/db6726bf-d6e9-43fa-828f-c78798598723-kube-api-access-2fh4c\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.532316 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-combined-ca-bundle\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.539328 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697b7cf9cc-qwr99"] Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.545532 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-config-data-custom\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.554585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-config-data\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.557537 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6t74\" (UniqueName: \"kubernetes.io/projected/048fa7a0-bb2f-4e32-b3c3-0c5af443caf9-kube-api-access-m6t74\") pod \"barbican-worker-5b9d486759-h8xx5\" (UID: \"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9\") " pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.607737 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6454f6bb5d-n6n86"] Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.609564 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.612416 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.624472 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6454f6bb5d-n6n86"] Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629041 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-config-data-custom\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-combined-ca-bundle\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-combined-ca-bundle\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6726bf-d6e9-43fa-828f-c78798598723-logs\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-dns-svc\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629187 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-sb\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629209 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24422fbf-e081-4b3c-895c-91f5cfe2a340-logs\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629230 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnx2q\" (UniqueName: \"kubernetes.io/projected/4fbf0c52-301f-4d29-a333-e8dc4a25448c-kube-api-access-bnx2q\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629253 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-config-data\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629273 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fh4c\" (UniqueName: \"kubernetes.io/projected/db6726bf-d6e9-43fa-828f-c78798598723-kube-api-access-2fh4c\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629305 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-config-data\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629327 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-nb\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629369 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-config-data-custom\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.629389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s64p\" (UniqueName: \"kubernetes.io/projected/24422fbf-e081-4b3c-895c-91f5cfe2a340-kube-api-access-9s64p\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.634350 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.635010 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-nb\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.635595 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-sb\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.635716 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-dns-svc\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.636101 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6726bf-d6e9-43fa-828f-c78798598723-logs\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.639803 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-config-data\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.639848 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-combined-ca-bundle\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.653780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnx2q\" (UniqueName: \"kubernetes.io/projected/4fbf0c52-301f-4d29-a333-e8dc4a25448c-kube-api-access-bnx2q\") pod \"dnsmasq-dns-697b7cf9cc-qwr99\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.654140 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db6726bf-d6e9-43fa-828f-c78798598723-config-data-custom\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.658199 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fh4c\" (UniqueName: \"kubernetes.io/projected/db6726bf-d6e9-43fa-828f-c78798598723-kube-api-access-2fh4c\") pod \"barbican-keystone-listener-78f54888c6-ht5kn\" (UID: \"db6726bf-d6e9-43fa-828f-c78798598723\") " pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.729258 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b9d486759-h8xx5" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.733032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-config-data\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.733107 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-config-data-custom\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.733135 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s64p\" (UniqueName: \"kubernetes.io/projected/24422fbf-e081-4b3c-895c-91f5cfe2a340-kube-api-access-9s64p\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.733205 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-combined-ca-bundle\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.733307 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24422fbf-e081-4b3c-895c-91f5cfe2a340-logs\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.733673 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24422fbf-e081-4b3c-895c-91f5cfe2a340-logs\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.740361 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-config-data-custom\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.741464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-config-data\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.742241 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24422fbf-e081-4b3c-895c-91f5cfe2a340-combined-ca-bundle\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.763282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s64p\" (UniqueName: \"kubernetes.io/projected/24422fbf-e081-4b3c-895c-91f5cfe2a340-kube-api-access-9s64p\") pod \"barbican-api-6454f6bb5d-n6n86\" (UID: \"24422fbf-e081-4b3c-895c-91f5cfe2a340\") " pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.770262 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sn7zk" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="registry-server" containerID="cri-o://fa2dc1ce8d93d2bd4aae915e0174dde1d0bd907bb94cf673b800fa843a298220" gracePeriod=2 Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.798182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" Dec 05 12:39:19 crc kubenswrapper[4809]: I1205 12:39:19.936240 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.015789 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.236022 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b9d486759-h8xx5"] Dec 05 12:39:20 crc kubenswrapper[4809]: W1205 12:39:20.435410 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fbf0c52_301f_4d29_a333_e8dc4a25448c.slice/crio-ef214ccc672b4bea54af3ec44475e9a10d193c081f8f9cf2dc41cbf871303b8c WatchSource:0}: Error finding container ef214ccc672b4bea54af3ec44475e9a10d193c081f8f9cf2dc41cbf871303b8c: Status 404 returned error can't find the container with id ef214ccc672b4bea54af3ec44475e9a10d193c081f8f9cf2dc41cbf871303b8c Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.439004 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697b7cf9cc-qwr99"] Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.520395 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-78f54888c6-ht5kn"] Dec 05 12:39:20 crc kubenswrapper[4809]: W1205 12:39:20.523758 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb6726bf_d6e9_43fa_828f_c78798598723.slice/crio-b35d9c2776869101ca40b749962129a2ce11b90fa5c6052fcc7d07e6a0964335 WatchSource:0}: Error finding container b35d9c2776869101ca40b749962129a2ce11b90fa5c6052fcc7d07e6a0964335: Status 404 returned error can't find the container with id b35d9c2776869101ca40b749962129a2ce11b90fa5c6052fcc7d07e6a0964335 Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.583052 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6454f6bb5d-n6n86"] Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.788035 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6454f6bb5d-n6n86" event={"ID":"24422fbf-e081-4b3c-895c-91f5cfe2a340","Type":"ContainerStarted","Data":"28ca3dd068c13449dd41e2a323ef613a543b880d0a14fd5039c6e93952000b9b"} Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.794185 4809 generic.go:334] "Generic (PLEG): container finished" podID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerID="fa2dc1ce8d93d2bd4aae915e0174dde1d0bd907bb94cf673b800fa843a298220" exitCode=0 Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.794242 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sn7zk" event={"ID":"be2dd974-bccc-4d5f-9450-780cb4e91b0d","Type":"ContainerDied","Data":"fa2dc1ce8d93d2bd4aae915e0174dde1d0bd907bb94cf673b800fa843a298220"} Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.796005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b9d486759-h8xx5" event={"ID":"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9","Type":"ContainerStarted","Data":"953c972a9e64bd0c192f1f6848344596aa7ffe5780eb0af8fef37471d64caf41"} Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.797491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" event={"ID":"4fbf0c52-301f-4d29-a333-e8dc4a25448c","Type":"ContainerStarted","Data":"ef214ccc672b4bea54af3ec44475e9a10d193c081f8f9cf2dc41cbf871303b8c"} Dec 05 12:39:20 crc kubenswrapper[4809]: I1205 12:39:20.802274 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" event={"ID":"db6726bf-d6e9-43fa-828f-c78798598723","Type":"ContainerStarted","Data":"b35d9c2776869101ca40b749962129a2ce11b90fa5c6052fcc7d07e6a0964335"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.015623 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.062015 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bt52\" (UniqueName: \"kubernetes.io/projected/be2dd974-bccc-4d5f-9450-780cb4e91b0d-kube-api-access-5bt52\") pod \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.062148 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-utilities\") pod \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.062281 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-catalog-content\") pod \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\" (UID: \"be2dd974-bccc-4d5f-9450-780cb4e91b0d\") " Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.065112 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-utilities" (OuterVolumeSpecName: "utilities") pod "be2dd974-bccc-4d5f-9450-780cb4e91b0d" (UID: "be2dd974-bccc-4d5f-9450-780cb4e91b0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.066228 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be2dd974-bccc-4d5f-9450-780cb4e91b0d-kube-api-access-5bt52" (OuterVolumeSpecName: "kube-api-access-5bt52") pod "be2dd974-bccc-4d5f-9450-780cb4e91b0d" (UID: "be2dd974-bccc-4d5f-9450-780cb4e91b0d"). InnerVolumeSpecName "kube-api-access-5bt52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.165583 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.165611 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bt52\" (UniqueName: \"kubernetes.io/projected/be2dd974-bccc-4d5f-9450-780cb4e91b0d-kube-api-access-5bt52\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.180511 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be2dd974-bccc-4d5f-9450-780cb4e91b0d" (UID: "be2dd974-bccc-4d5f-9450-780cb4e91b0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.266739 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2dd974-bccc-4d5f-9450-780cb4e91b0d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.821083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sn7zk" event={"ID":"be2dd974-bccc-4d5f-9450-780cb4e91b0d","Type":"ContainerDied","Data":"d61b051ce2a2b78f08892192810c774687872e6a4e8b9e12b5ea9ad4965d24b8"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.821153 4809 scope.go:117] "RemoveContainer" containerID="fa2dc1ce8d93d2bd4aae915e0174dde1d0bd907bb94cf673b800fa843a298220" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.821171 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sn7zk" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.823993 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b9d486759-h8xx5" event={"ID":"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9","Type":"ContainerStarted","Data":"a47c04ba73bf61aa254f5f9a0e3da445b519a1d7f63bb6fd1b35f847a0446ee6"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.824047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b9d486759-h8xx5" event={"ID":"048fa7a0-bb2f-4e32-b3c3-0c5af443caf9","Type":"ContainerStarted","Data":"7019832b0d17e7f64f2d0bf9c10a2a3e5fcf277a738d923bf6c0155b78344ea3"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.827703 4809 generic.go:334] "Generic (PLEG): container finished" podID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerID="09171c2d9eccca7f147ed9f35aafa857991474b6a5d157f5eb9154cf262b7af7" exitCode=0 Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.827780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" event={"ID":"4fbf0c52-301f-4d29-a333-e8dc4a25448c","Type":"ContainerDied","Data":"09171c2d9eccca7f147ed9f35aafa857991474b6a5d157f5eb9154cf262b7af7"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.830693 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" event={"ID":"db6726bf-d6e9-43fa-828f-c78798598723","Type":"ContainerStarted","Data":"8ba16475c30e0fc320e7b2f136a482ab2781875a63137623a3e7ca33b27b1b26"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.830727 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" event={"ID":"db6726bf-d6e9-43fa-828f-c78798598723","Type":"ContainerStarted","Data":"e0d677a073a89424f95a231f90b26cfd0c2156ce72ac9c7b96d54e2e0234d935"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.832642 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6454f6bb5d-n6n86" event={"ID":"24422fbf-e081-4b3c-895c-91f5cfe2a340","Type":"ContainerStarted","Data":"c8701de21ab8cba115e2927184389d4967546179ba640b59bfd81dffd51de40a"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.832688 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6454f6bb5d-n6n86" event={"ID":"24422fbf-e081-4b3c-895c-91f5cfe2a340","Type":"ContainerStarted","Data":"d2c60c781ac7e3cd299d7f5d6767f0175773076a0f26ae51cef79d834646d4fd"} Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.849893 4809 scope.go:117] "RemoveContainer" containerID="55834004d1931f88ba4ff456fd01b4d00d7af0d82bef62f34150f955105f29aa" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.884830 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5b9d486759-h8xx5" podStartSLOduration=2.8844374999999998 podStartE2EDuration="2.8844375s" podCreationTimestamp="2025-12-05 12:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:39:21.850144233 +0000 UTC m=+5457.241120791" watchObservedRunningTime="2025-12-05 12:39:21.8844375 +0000 UTC m=+5457.275414058" Dec 05 12:39:21 crc kubenswrapper[4809]: I1205 12:39:21.945969 4809 scope.go:117] "RemoveContainer" containerID="5089cae48dfdd2513ce14536089cf8bfb4921bf78f1ee72639f94a7e3173504a" Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.044522 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sn7zk"] Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.052934 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sn7zk"] Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.843960 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" event={"ID":"4fbf0c52-301f-4d29-a333-e8dc4a25448c","Type":"ContainerStarted","Data":"f336e1d363bde633a5587d6a97ae454e2a99c824f9f765c91e0f445d4872b5f8"} Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.844106 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.845569 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.845735 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.864134 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" podStartSLOduration=3.864112224 podStartE2EDuration="3.864112224s" podCreationTimestamp="2025-12-05 12:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:39:22.862705066 +0000 UTC m=+5458.253681644" watchObservedRunningTime="2025-12-05 12:39:22.864112224 +0000 UTC m=+5458.255088782" Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.881982 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-78f54888c6-ht5kn" podStartSLOduration=3.881964542 podStartE2EDuration="3.881964542s" podCreationTimestamp="2025-12-05 12:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:39:22.879753453 +0000 UTC m=+5458.270730011" watchObservedRunningTime="2025-12-05 12:39:22.881964542 +0000 UTC m=+5458.272941100" Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.885312 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" path="/var/lib/kubelet/pods/be2dd974-bccc-4d5f-9450-780cb4e91b0d/volumes" Dec 05 12:39:22 crc kubenswrapper[4809]: I1205 12:39:22.908071 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6454f6bb5d-n6n86" podStartSLOduration=3.908049029 podStartE2EDuration="3.908049029s" podCreationTimestamp="2025-12-05 12:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:39:22.898611587 +0000 UTC m=+5458.289588165" watchObservedRunningTime="2025-12-05 12:39:22.908049029 +0000 UTC m=+5458.299025587" Dec 05 12:39:29 crc kubenswrapper[4809]: I1205 12:39:29.938859 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:39:29 crc kubenswrapper[4809]: I1205 12:39:29.993195 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk"] Dec 05 12:39:29 crc kubenswrapper[4809]: I1205 12:39:29.993483 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerName="dnsmasq-dns" containerID="cri-o://00390f625e401af7c7857b5734a563c4b851ff0809da3e2a56b8911a8191eaa1" gracePeriod=10 Dec 05 12:39:30 crc kubenswrapper[4809]: I1205 12:39:30.879646 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.19:5353: connect: connection refused" Dec 05 12:39:31 crc kubenswrapper[4809]: I1205 12:39:31.505134 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:31 crc kubenswrapper[4809]: I1205 12:39:31.622033 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6454f6bb5d-n6n86" Dec 05 12:39:32 crc kubenswrapper[4809]: I1205 12:39:32.970853 4809 generic.go:334] "Generic (PLEG): container finished" podID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerID="00390f625e401af7c7857b5734a563c4b851ff0809da3e2a56b8911a8191eaa1" exitCode=0 Dec 05 12:39:32 crc kubenswrapper[4809]: I1205 12:39:32.970906 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" event={"ID":"99018d5d-99c4-458c-b322-f650bcdeb04b","Type":"ContainerDied","Data":"00390f625e401af7c7857b5734a563c4b851ff0809da3e2a56b8911a8191eaa1"} Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.607782 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.757425 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-config\") pod \"99018d5d-99c4-458c-b322-f650bcdeb04b\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.757564 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-nb\") pod \"99018d5d-99c4-458c-b322-f650bcdeb04b\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.757587 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-sb\") pod \"99018d5d-99c4-458c-b322-f650bcdeb04b\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.757713 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w2vb\" (UniqueName: \"kubernetes.io/projected/99018d5d-99c4-458c-b322-f650bcdeb04b-kube-api-access-5w2vb\") pod \"99018d5d-99c4-458c-b322-f650bcdeb04b\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.758824 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-dns-svc\") pod \"99018d5d-99c4-458c-b322-f650bcdeb04b\" (UID: \"99018d5d-99c4-458c-b322-f650bcdeb04b\") " Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.763756 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99018d5d-99c4-458c-b322-f650bcdeb04b-kube-api-access-5w2vb" (OuterVolumeSpecName: "kube-api-access-5w2vb") pod "99018d5d-99c4-458c-b322-f650bcdeb04b" (UID: "99018d5d-99c4-458c-b322-f650bcdeb04b"). InnerVolumeSpecName "kube-api-access-5w2vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.799118 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "99018d5d-99c4-458c-b322-f650bcdeb04b" (UID: "99018d5d-99c4-458c-b322-f650bcdeb04b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.800619 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99018d5d-99c4-458c-b322-f650bcdeb04b" (UID: "99018d5d-99c4-458c-b322-f650bcdeb04b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.802132 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-config" (OuterVolumeSpecName: "config") pod "99018d5d-99c4-458c-b322-f650bcdeb04b" (UID: "99018d5d-99c4-458c-b322-f650bcdeb04b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.807100 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "99018d5d-99c4-458c-b322-f650bcdeb04b" (UID: "99018d5d-99c4-458c-b322-f650bcdeb04b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.861090 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.861245 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.861370 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.861506 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99018d5d-99c4-458c-b322-f650bcdeb04b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.861715 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w2vb\" (UniqueName: \"kubernetes.io/projected/99018d5d-99c4-458c-b322-f650bcdeb04b-kube-api-access-5w2vb\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.989528 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" event={"ID":"99018d5d-99c4-458c-b322-f650bcdeb04b","Type":"ContainerDied","Data":"3cbd6c67f2df918bd0bc1b354d9731f8763c24a3368fad007237121636d2d013"} Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.989615 4809 scope.go:117] "RemoveContainer" containerID="00390f625e401af7c7857b5734a563c4b851ff0809da3e2a56b8911a8191eaa1" Dec 05 12:39:34 crc kubenswrapper[4809]: I1205 12:39:34.989951 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk" Dec 05 12:39:35 crc kubenswrapper[4809]: I1205 12:39:35.015549 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk"] Dec 05 12:39:35 crc kubenswrapper[4809]: I1205 12:39:35.017246 4809 scope.go:117] "RemoveContainer" containerID="d772570f4d19ee95279fe102fb66a51c2c4823a98b5fc0669ce187179bd79477" Dec 05 12:39:35 crc kubenswrapper[4809]: I1205 12:39:35.024924 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c6bcd7cbf-9dfpk"] Dec 05 12:39:36 crc kubenswrapper[4809]: I1205 12:39:36.881055 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" path="/var/lib/kubelet/pods/99018d5d-99c4-458c-b322-f650bcdeb04b/volumes" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.234871 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dm5zl"] Dec 05 12:39:37 crc kubenswrapper[4809]: E1205 12:39:37.235259 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="registry-server" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.235284 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="registry-server" Dec 05 12:39:37 crc kubenswrapper[4809]: E1205 12:39:37.235301 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="extract-content" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.235310 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="extract-content" Dec 05 12:39:37 crc kubenswrapper[4809]: E1205 12:39:37.235325 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerName="init" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.235332 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerName="init" Dec 05 12:39:37 crc kubenswrapper[4809]: E1205 12:39:37.235345 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="extract-utilities" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.235353 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="extract-utilities" Dec 05 12:39:37 crc kubenswrapper[4809]: E1205 12:39:37.235366 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerName="dnsmasq-dns" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.235375 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerName="dnsmasq-dns" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.235548 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="99018d5d-99c4-458c-b322-f650bcdeb04b" containerName="dnsmasq-dns" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.235572 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="be2dd974-bccc-4d5f-9450-780cb4e91b0d" containerName="registry-server" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.237332 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.245268 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm5zl"] Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.410285 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-catalog-content\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.410548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-utilities\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.410585 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgppp\" (UniqueName: \"kubernetes.io/projected/7efc795d-5f4e-42a3-a38b-77a74dc124ed-kube-api-access-pgppp\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.512760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-catalog-content\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.512830 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-utilities\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.512866 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgppp\" (UniqueName: \"kubernetes.io/projected/7efc795d-5f4e-42a3-a38b-77a74dc124ed-kube-api-access-pgppp\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.513284 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-utilities\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.513321 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-catalog-content\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.535524 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgppp\" (UniqueName: \"kubernetes.io/projected/7efc795d-5f4e-42a3-a38b-77a74dc124ed-kube-api-access-pgppp\") pod \"redhat-marketplace-dm5zl\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:37 crc kubenswrapper[4809]: I1205 12:39:37.560601 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:38 crc kubenswrapper[4809]: I1205 12:39:38.047224 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm5zl"] Dec 05 12:39:39 crc kubenswrapper[4809]: I1205 12:39:39.031706 4809 generic.go:334] "Generic (PLEG): container finished" podID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerID="d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c" exitCode=0 Dec 05 12:39:39 crc kubenswrapper[4809]: I1205 12:39:39.031808 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm5zl" event={"ID":"7efc795d-5f4e-42a3-a38b-77a74dc124ed","Type":"ContainerDied","Data":"d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c"} Dec 05 12:39:39 crc kubenswrapper[4809]: I1205 12:39:39.032109 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm5zl" event={"ID":"7efc795d-5f4e-42a3-a38b-77a74dc124ed","Type":"ContainerStarted","Data":"5dcb39d2d74eed270ff4941effad85a7c106acf9f429abde31b2de50146c3e04"} Dec 05 12:39:40 crc kubenswrapper[4809]: I1205 12:39:40.041747 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm5zl" event={"ID":"7efc795d-5f4e-42a3-a38b-77a74dc124ed","Type":"ContainerStarted","Data":"9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8"} Dec 05 12:39:41 crc kubenswrapper[4809]: I1205 12:39:41.050326 4809 generic.go:334] "Generic (PLEG): container finished" podID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerID="9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8" exitCode=0 Dec 05 12:39:41 crc kubenswrapper[4809]: I1205 12:39:41.050509 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm5zl" event={"ID":"7efc795d-5f4e-42a3-a38b-77a74dc124ed","Type":"ContainerDied","Data":"9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8"} Dec 05 12:39:42 crc kubenswrapper[4809]: I1205 12:39:42.060443 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm5zl" event={"ID":"7efc795d-5f4e-42a3-a38b-77a74dc124ed","Type":"ContainerStarted","Data":"bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe"} Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.836856 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dm5zl" podStartSLOduration=4.246793238 podStartE2EDuration="6.836833957s" podCreationTimestamp="2025-12-05 12:39:37 +0000 UTC" firstStartedPulling="2025-12-05 12:39:39.033569358 +0000 UTC m=+5474.424545936" lastFinishedPulling="2025-12-05 12:39:41.623610097 +0000 UTC m=+5477.014586655" observedRunningTime="2025-12-05 12:39:42.082120071 +0000 UTC m=+5477.473096629" watchObservedRunningTime="2025-12-05 12:39:43.836833957 +0000 UTC m=+5479.227810515" Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.844681 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-lgwdk"] Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.846003 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.854521 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lgwdk"] Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.941165 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e34c-account-create-update-vffl5"] Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.943593 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.946025 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 05 12:39:43 crc kubenswrapper[4809]: I1205 12:39:43.954621 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e34c-account-create-update-vffl5"] Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.031071 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb1ffcc3-7e44-400b-a3c8-65071612f16f-operator-scripts\") pod \"neutron-db-create-lgwdk\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.031151 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6xk7\" (UniqueName: \"kubernetes.io/projected/cb1ffcc3-7e44-400b-a3c8-65071612f16f-kube-api-access-d6xk7\") pod \"neutron-db-create-lgwdk\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.031207 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-operator-scripts\") pod \"neutron-e34c-account-create-update-vffl5\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.031254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbjc5\" (UniqueName: \"kubernetes.io/projected/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-kube-api-access-xbjc5\") pod \"neutron-e34c-account-create-update-vffl5\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.046870 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.046929 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.132667 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb1ffcc3-7e44-400b-a3c8-65071612f16f-operator-scripts\") pod \"neutron-db-create-lgwdk\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.132785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6xk7\" (UniqueName: \"kubernetes.io/projected/cb1ffcc3-7e44-400b-a3c8-65071612f16f-kube-api-access-d6xk7\") pod \"neutron-db-create-lgwdk\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.132812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-operator-scripts\") pod \"neutron-e34c-account-create-update-vffl5\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.132864 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbjc5\" (UniqueName: \"kubernetes.io/projected/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-kube-api-access-xbjc5\") pod \"neutron-e34c-account-create-update-vffl5\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.133461 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb1ffcc3-7e44-400b-a3c8-65071612f16f-operator-scripts\") pod \"neutron-db-create-lgwdk\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.133597 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-operator-scripts\") pod \"neutron-e34c-account-create-update-vffl5\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.151660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbjc5\" (UniqueName: \"kubernetes.io/projected/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-kube-api-access-xbjc5\") pod \"neutron-e34c-account-create-update-vffl5\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.152555 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6xk7\" (UniqueName: \"kubernetes.io/projected/cb1ffcc3-7e44-400b-a3c8-65071612f16f-kube-api-access-d6xk7\") pod \"neutron-db-create-lgwdk\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.169095 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.264931 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.650023 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lgwdk"] Dec 05 12:39:44 crc kubenswrapper[4809]: W1205 12:39:44.650305 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb1ffcc3_7e44_400b_a3c8_65071612f16f.slice/crio-194e349a0d94291530669b74547acc7321c562720a31efcdfcbb40a0bdeb5582 WatchSource:0}: Error finding container 194e349a0d94291530669b74547acc7321c562720a31efcdfcbb40a0bdeb5582: Status 404 returned error can't find the container with id 194e349a0d94291530669b74547acc7321c562720a31efcdfcbb40a0bdeb5582 Dec 05 12:39:44 crc kubenswrapper[4809]: I1205 12:39:44.739716 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e34c-account-create-update-vffl5"] Dec 05 12:39:44 crc kubenswrapper[4809]: W1205 12:39:44.747348 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccc4bb97_a08f_40a3_b8f6_f014099c04cc.slice/crio-f47eb7fd61caf3a99f771ef626828962dee2a81942ac84492dbab15144b1b1e0 WatchSource:0}: Error finding container f47eb7fd61caf3a99f771ef626828962dee2a81942ac84492dbab15144b1b1e0: Status 404 returned error can't find the container with id f47eb7fd61caf3a99f771ef626828962dee2a81942ac84492dbab15144b1b1e0 Dec 05 12:39:45 crc kubenswrapper[4809]: I1205 12:39:45.087100 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e34c-account-create-update-vffl5" event={"ID":"ccc4bb97-a08f-40a3-b8f6-f014099c04cc","Type":"ContainerStarted","Data":"f47eb7fd61caf3a99f771ef626828962dee2a81942ac84492dbab15144b1b1e0"} Dec 05 12:39:45 crc kubenswrapper[4809]: I1205 12:39:45.088513 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lgwdk" event={"ID":"cb1ffcc3-7e44-400b-a3c8-65071612f16f","Type":"ContainerStarted","Data":"194e349a0d94291530669b74547acc7321c562720a31efcdfcbb40a0bdeb5582"} Dec 05 12:39:47 crc kubenswrapper[4809]: I1205 12:39:47.106602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e34c-account-create-update-vffl5" event={"ID":"ccc4bb97-a08f-40a3-b8f6-f014099c04cc","Type":"ContainerStarted","Data":"853062e39ea30216ee76ee1fada3f2dad4b363851250da8dc37db47099fafcee"} Dec 05 12:39:47 crc kubenswrapper[4809]: I1205 12:39:47.561672 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:47 crc kubenswrapper[4809]: I1205 12:39:47.561722 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:47 crc kubenswrapper[4809]: I1205 12:39:47.603214 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:48 crc kubenswrapper[4809]: I1205 12:39:48.116169 4809 generic.go:334] "Generic (PLEG): container finished" podID="ccc4bb97-a08f-40a3-b8f6-f014099c04cc" containerID="853062e39ea30216ee76ee1fada3f2dad4b363851250da8dc37db47099fafcee" exitCode=0 Dec 05 12:39:48 crc kubenswrapper[4809]: I1205 12:39:48.116211 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e34c-account-create-update-vffl5" event={"ID":"ccc4bb97-a08f-40a3-b8f6-f014099c04cc","Type":"ContainerDied","Data":"853062e39ea30216ee76ee1fada3f2dad4b363851250da8dc37db47099fafcee"} Dec 05 12:39:48 crc kubenswrapper[4809]: I1205 12:39:48.117999 4809 generic.go:334] "Generic (PLEG): container finished" podID="cb1ffcc3-7e44-400b-a3c8-65071612f16f" containerID="f3bfe3273ad663d02342cfa66f8aeb7369345b1e3287267477fc080ecfdf51c6" exitCode=0 Dec 05 12:39:48 crc kubenswrapper[4809]: I1205 12:39:48.118047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lgwdk" event={"ID":"cb1ffcc3-7e44-400b-a3c8-65071612f16f","Type":"ContainerDied","Data":"f3bfe3273ad663d02342cfa66f8aeb7369345b1e3287267477fc080ecfdf51c6"} Dec 05 12:39:48 crc kubenswrapper[4809]: I1205 12:39:48.179849 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.568991 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.582428 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.737578 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6xk7\" (UniqueName: \"kubernetes.io/projected/cb1ffcc3-7e44-400b-a3c8-65071612f16f-kube-api-access-d6xk7\") pod \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.737757 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbjc5\" (UniqueName: \"kubernetes.io/projected/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-kube-api-access-xbjc5\") pod \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.737835 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb1ffcc3-7e44-400b-a3c8-65071612f16f-operator-scripts\") pod \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\" (UID: \"cb1ffcc3-7e44-400b-a3c8-65071612f16f\") " Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.737923 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-operator-scripts\") pod \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\" (UID: \"ccc4bb97-a08f-40a3-b8f6-f014099c04cc\") " Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.738838 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ccc4bb97-a08f-40a3-b8f6-f014099c04cc" (UID: "ccc4bb97-a08f-40a3-b8f6-f014099c04cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.738838 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb1ffcc3-7e44-400b-a3c8-65071612f16f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cb1ffcc3-7e44-400b-a3c8-65071612f16f" (UID: "cb1ffcc3-7e44-400b-a3c8-65071612f16f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.744476 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb1ffcc3-7e44-400b-a3c8-65071612f16f-kube-api-access-d6xk7" (OuterVolumeSpecName: "kube-api-access-d6xk7") pod "cb1ffcc3-7e44-400b-a3c8-65071612f16f" (UID: "cb1ffcc3-7e44-400b-a3c8-65071612f16f"). InnerVolumeSpecName "kube-api-access-d6xk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.744559 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-kube-api-access-xbjc5" (OuterVolumeSpecName: "kube-api-access-xbjc5") pod "ccc4bb97-a08f-40a3-b8f6-f014099c04cc" (UID: "ccc4bb97-a08f-40a3-b8f6-f014099c04cc"). InnerVolumeSpecName "kube-api-access-xbjc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.839946 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.839983 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6xk7\" (UniqueName: \"kubernetes.io/projected/cb1ffcc3-7e44-400b-a3c8-65071612f16f-kube-api-access-d6xk7\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.839996 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbjc5\" (UniqueName: \"kubernetes.io/projected/ccc4bb97-a08f-40a3-b8f6-f014099c04cc-kube-api-access-xbjc5\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.840007 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb1ffcc3-7e44-400b-a3c8-65071612f16f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:49 crc kubenswrapper[4809]: I1205 12:39:49.904679 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm5zl"] Dec 05 12:39:50 crc kubenswrapper[4809]: I1205 12:39:50.154424 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e34c-account-create-update-vffl5" Dec 05 12:39:50 crc kubenswrapper[4809]: I1205 12:39:50.154418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e34c-account-create-update-vffl5" event={"ID":"ccc4bb97-a08f-40a3-b8f6-f014099c04cc","Type":"ContainerDied","Data":"f47eb7fd61caf3a99f771ef626828962dee2a81942ac84492dbab15144b1b1e0"} Dec 05 12:39:50 crc kubenswrapper[4809]: I1205 12:39:50.154937 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f47eb7fd61caf3a99f771ef626828962dee2a81942ac84492dbab15144b1b1e0" Dec 05 12:39:50 crc kubenswrapper[4809]: I1205 12:39:50.156861 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lgwdk" event={"ID":"cb1ffcc3-7e44-400b-a3c8-65071612f16f","Type":"ContainerDied","Data":"194e349a0d94291530669b74547acc7321c562720a31efcdfcbb40a0bdeb5582"} Dec 05 12:39:50 crc kubenswrapper[4809]: I1205 12:39:50.156887 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="194e349a0d94291530669b74547acc7321c562720a31efcdfcbb40a0bdeb5582" Dec 05 12:39:50 crc kubenswrapper[4809]: I1205 12:39:50.156890 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lgwdk" Dec 05 12:39:50 crc kubenswrapper[4809]: I1205 12:39:50.157135 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dm5zl" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="registry-server" containerID="cri-o://bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe" gracePeriod=2 Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.006762 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.081564 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-catalog-content\") pod \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.081666 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-utilities\") pod \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.081780 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgppp\" (UniqueName: \"kubernetes.io/projected/7efc795d-5f4e-42a3-a38b-77a74dc124ed-kube-api-access-pgppp\") pod \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\" (UID: \"7efc795d-5f4e-42a3-a38b-77a74dc124ed\") " Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.083981 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-utilities" (OuterVolumeSpecName: "utilities") pod "7efc795d-5f4e-42a3-a38b-77a74dc124ed" (UID: "7efc795d-5f4e-42a3-a38b-77a74dc124ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.087437 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7efc795d-5f4e-42a3-a38b-77a74dc124ed-kube-api-access-pgppp" (OuterVolumeSpecName: "kube-api-access-pgppp") pod "7efc795d-5f4e-42a3-a38b-77a74dc124ed" (UID: "7efc795d-5f4e-42a3-a38b-77a74dc124ed"). InnerVolumeSpecName "kube-api-access-pgppp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.104953 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7efc795d-5f4e-42a3-a38b-77a74dc124ed" (UID: "7efc795d-5f4e-42a3-a38b-77a74dc124ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.175191 4809 generic.go:334] "Generic (PLEG): container finished" podID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerID="bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe" exitCode=0 Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.175230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm5zl" event={"ID":"7efc795d-5f4e-42a3-a38b-77a74dc124ed","Type":"ContainerDied","Data":"bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe"} Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.175262 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm5zl" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.175278 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm5zl" event={"ID":"7efc795d-5f4e-42a3-a38b-77a74dc124ed","Type":"ContainerDied","Data":"5dcb39d2d74eed270ff4941effad85a7c106acf9f429abde31b2de50146c3e04"} Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.175296 4809 scope.go:117] "RemoveContainer" containerID="bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.183802 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgppp\" (UniqueName: \"kubernetes.io/projected/7efc795d-5f4e-42a3-a38b-77a74dc124ed-kube-api-access-pgppp\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.184096 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.184109 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7efc795d-5f4e-42a3-a38b-77a74dc124ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.207967 4809 scope.go:117] "RemoveContainer" containerID="9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.210289 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm5zl"] Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.217957 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm5zl"] Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.239757 4809 scope.go:117] "RemoveContainer" containerID="d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.261466 4809 scope.go:117] "RemoveContainer" containerID="bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe" Dec 05 12:39:52 crc kubenswrapper[4809]: E1205 12:39:52.261941 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe\": container with ID starting with bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe not found: ID does not exist" containerID="bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.261995 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe"} err="failed to get container status \"bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe\": rpc error: code = NotFound desc = could not find container \"bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe\": container with ID starting with bff39c25818d67b3c481b3aeca545e1efdd1e518e6e429e6b344276f27577ffe not found: ID does not exist" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.262025 4809 scope.go:117] "RemoveContainer" containerID="9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8" Dec 05 12:39:52 crc kubenswrapper[4809]: E1205 12:39:52.262384 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8\": container with ID starting with 9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8 not found: ID does not exist" containerID="9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.262423 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8"} err="failed to get container status \"9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8\": rpc error: code = NotFound desc = could not find container \"9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8\": container with ID starting with 9653b97f855d9e9c0abd43397f72d5712de6ad83c32ea00d6a1f67e4edfd8dc8 not found: ID does not exist" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.262475 4809 scope.go:117] "RemoveContainer" containerID="d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c" Dec 05 12:39:52 crc kubenswrapper[4809]: E1205 12:39:52.262875 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c\": container with ID starting with d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c not found: ID does not exist" containerID="d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.262907 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c"} err="failed to get container status \"d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c\": rpc error: code = NotFound desc = could not find container \"d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c\": container with ID starting with d0e70a8b3ed69a819f8c86d895160099a3aa5409c0171a12b378f7090ba8e04c not found: ID does not exist" Dec 05 12:39:52 crc kubenswrapper[4809]: I1205 12:39:52.921658 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" path="/var/lib/kubelet/pods/7efc795d-5f4e-42a3-a38b-77a74dc124ed/volumes" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.281497 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-nxz6q"] Dec 05 12:39:54 crc kubenswrapper[4809]: E1205 12:39:54.282112 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc4bb97-a08f-40a3-b8f6-f014099c04cc" containerName="mariadb-account-create-update" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282125 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc4bb97-a08f-40a3-b8f6-f014099c04cc" containerName="mariadb-account-create-update" Dec 05 12:39:54 crc kubenswrapper[4809]: E1205 12:39:54.282152 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="extract-utilities" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282159 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="extract-utilities" Dec 05 12:39:54 crc kubenswrapper[4809]: E1205 12:39:54.282173 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1ffcc3-7e44-400b-a3c8-65071612f16f" containerName="mariadb-database-create" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282180 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1ffcc3-7e44-400b-a3c8-65071612f16f" containerName="mariadb-database-create" Dec 05 12:39:54 crc kubenswrapper[4809]: E1205 12:39:54.282197 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="extract-content" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282203 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="extract-content" Dec 05 12:39:54 crc kubenswrapper[4809]: E1205 12:39:54.282226 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="registry-server" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282233 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="registry-server" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282451 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb1ffcc3-7e44-400b-a3c8-65071612f16f" containerName="mariadb-database-create" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282468 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc4bb97-a08f-40a3-b8f6-f014099c04cc" containerName="mariadb-account-create-update" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.282492 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7efc795d-5f4e-42a3-a38b-77a74dc124ed" containerName="registry-server" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.285103 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.287771 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.288165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.288267 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-sxtz6" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.304547 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nxz6q"] Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.435235 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v27cw\" (UniqueName: \"kubernetes.io/projected/67f36fae-3a5c-400e-afd9-1c3c33598973-kube-api-access-v27cw\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.435319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-combined-ca-bundle\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.435824 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-config\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.538625 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-config\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.538730 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v27cw\" (UniqueName: \"kubernetes.io/projected/67f36fae-3a5c-400e-afd9-1c3c33598973-kube-api-access-v27cw\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.538796 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-combined-ca-bundle\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.544098 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-combined-ca-bundle\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.544850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-config\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.570037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v27cw\" (UniqueName: \"kubernetes.io/projected/67f36fae-3a5c-400e-afd9-1c3c33598973-kube-api-access-v27cw\") pod \"neutron-db-sync-nxz6q\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:54 crc kubenswrapper[4809]: I1205 12:39:54.657109 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:39:55 crc kubenswrapper[4809]: I1205 12:39:55.072851 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nxz6q"] Dec 05 12:39:55 crc kubenswrapper[4809]: I1205 12:39:55.200604 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nxz6q" event={"ID":"67f36fae-3a5c-400e-afd9-1c3c33598973","Type":"ContainerStarted","Data":"cef5aea70fe4e0239d77ea31ffae0fd1a40cbed97516d950198669c5f342b0ef"} Dec 05 12:39:56 crc kubenswrapper[4809]: I1205 12:39:56.209866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nxz6q" event={"ID":"67f36fae-3a5c-400e-afd9-1c3c33598973","Type":"ContainerStarted","Data":"3b4c71e457b2f63fa20300186a49d3267b0cb88c80c27fa2d12936932574c81a"} Dec 05 12:39:56 crc kubenswrapper[4809]: I1205 12:39:56.224049 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-nxz6q" podStartSLOduration=2.224024272 podStartE2EDuration="2.224024272s" podCreationTimestamp="2025-12-05 12:39:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:39:56.222784288 +0000 UTC m=+5491.613760846" watchObservedRunningTime="2025-12-05 12:39:56.224024272 +0000 UTC m=+5491.615000830" Dec 05 12:40:01 crc kubenswrapper[4809]: I1205 12:40:01.252964 4809 generic.go:334] "Generic (PLEG): container finished" podID="67f36fae-3a5c-400e-afd9-1c3c33598973" containerID="3b4c71e457b2f63fa20300186a49d3267b0cb88c80c27fa2d12936932574c81a" exitCode=0 Dec 05 12:40:01 crc kubenswrapper[4809]: I1205 12:40:01.253072 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nxz6q" event={"ID":"67f36fae-3a5c-400e-afd9-1c3c33598973","Type":"ContainerDied","Data":"3b4c71e457b2f63fa20300186a49d3267b0cb88c80c27fa2d12936932574c81a"} Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.559940 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.732382 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-config\") pod \"67f36fae-3a5c-400e-afd9-1c3c33598973\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.732804 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-combined-ca-bundle\") pod \"67f36fae-3a5c-400e-afd9-1c3c33598973\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.732840 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v27cw\" (UniqueName: \"kubernetes.io/projected/67f36fae-3a5c-400e-afd9-1c3c33598973-kube-api-access-v27cw\") pod \"67f36fae-3a5c-400e-afd9-1c3c33598973\" (UID: \"67f36fae-3a5c-400e-afd9-1c3c33598973\") " Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.738231 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67f36fae-3a5c-400e-afd9-1c3c33598973-kube-api-access-v27cw" (OuterVolumeSpecName: "kube-api-access-v27cw") pod "67f36fae-3a5c-400e-afd9-1c3c33598973" (UID: "67f36fae-3a5c-400e-afd9-1c3c33598973"). InnerVolumeSpecName "kube-api-access-v27cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.760556 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67f36fae-3a5c-400e-afd9-1c3c33598973" (UID: "67f36fae-3a5c-400e-afd9-1c3c33598973"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.760734 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-config" (OuterVolumeSpecName: "config") pod "67f36fae-3a5c-400e-afd9-1c3c33598973" (UID: "67f36fae-3a5c-400e-afd9-1c3c33598973"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.834915 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v27cw\" (UniqueName: \"kubernetes.io/projected/67f36fae-3a5c-400e-afd9-1c3c33598973-kube-api-access-v27cw\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.834950 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:02 crc kubenswrapper[4809]: I1205 12:40:02.834961 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67f36fae-3a5c-400e-afd9-1c3c33598973-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.270468 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nxz6q" event={"ID":"67f36fae-3a5c-400e-afd9-1c3c33598973","Type":"ContainerDied","Data":"cef5aea70fe4e0239d77ea31ffae0fd1a40cbed97516d950198669c5f342b0ef"} Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.270522 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cef5aea70fe4e0239d77ea31ffae0fd1a40cbed97516d950198669c5f342b0ef" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.270528 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nxz6q" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.489976 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84c6c75d89-m7nqm"] Dec 05 12:40:03 crc kubenswrapper[4809]: E1205 12:40:03.490392 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67f36fae-3a5c-400e-afd9-1c3c33598973" containerName="neutron-db-sync" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.490415 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="67f36fae-3a5c-400e-afd9-1c3c33598973" containerName="neutron-db-sync" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.490596 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="67f36fae-3a5c-400e-afd9-1c3c33598973" containerName="neutron-db-sync" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.492150 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.510761 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84c6c75d89-m7nqm"] Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.611034 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cd8c99689-xmlvs"] Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.612800 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.616186 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.616802 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.616963 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-sxtz6" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.622931 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cd8c99689-xmlvs"] Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.652923 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-config\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.653122 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-dns-svc\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.653277 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-nb\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.653316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4sb9\" (UniqueName: \"kubernetes.io/projected/f87668ca-d015-4ac3-b074-a49b7fab8991-kube-api-access-n4sb9\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.653919 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-sb\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.755907 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-combined-ca-bundle\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.756292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-nb\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.756331 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4sb9\" (UniqueName: \"kubernetes.io/projected/f87668ca-d015-4ac3-b074-a49b7fab8991-kube-api-access-n4sb9\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.756371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-sb\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.756406 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-httpd-config\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.756438 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-config\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.756460 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-config\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.756832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzrm2\" (UniqueName: \"kubernetes.io/projected/08b88df2-c726-4578-9f31-ccf6690e58a1-kube-api-access-zzrm2\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.757008 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-dns-svc\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.758290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-sb\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.758333 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-dns-svc\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.758357 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-config\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.758719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-nb\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.785410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4sb9\" (UniqueName: \"kubernetes.io/projected/f87668ca-d015-4ac3-b074-a49b7fab8991-kube-api-access-n4sb9\") pod \"dnsmasq-dns-84c6c75d89-m7nqm\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.811822 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.858834 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-httpd-config\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.858894 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-config\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.858938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzrm2\" (UniqueName: \"kubernetes.io/projected/08b88df2-c726-4578-9f31-ccf6690e58a1-kube-api-access-zzrm2\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.858984 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-combined-ca-bundle\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.864573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-httpd-config\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.867549 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-combined-ca-bundle\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.870915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/08b88df2-c726-4578-9f31-ccf6690e58a1-config\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.887410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzrm2\" (UniqueName: \"kubernetes.io/projected/08b88df2-c726-4578-9f31-ccf6690e58a1-kube-api-access-zzrm2\") pod \"neutron-6cd8c99689-xmlvs\" (UID: \"08b88df2-c726-4578-9f31-ccf6690e58a1\") " pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:03 crc kubenswrapper[4809]: I1205 12:40:03.933064 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:04 crc kubenswrapper[4809]: I1205 12:40:04.441711 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84c6c75d89-m7nqm"] Dec 05 12:40:04 crc kubenswrapper[4809]: I1205 12:40:04.510139 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cd8c99689-xmlvs"] Dec 05 12:40:04 crc kubenswrapper[4809]: W1205 12:40:04.512233 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08b88df2_c726_4578_9f31_ccf6690e58a1.slice/crio-c6b88266f3769048b7519299b7481b28766d98f699980ef9d5a33673ff972c86 WatchSource:0}: Error finding container c6b88266f3769048b7519299b7481b28766d98f699980ef9d5a33673ff972c86: Status 404 returned error can't find the container with id c6b88266f3769048b7519299b7481b28766d98f699980ef9d5a33673ff972c86 Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.295318 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cd8c99689-xmlvs" event={"ID":"08b88df2-c726-4578-9f31-ccf6690e58a1","Type":"ContainerStarted","Data":"25f17802fee926e4cce9c92f36b7d2de1c8ab0ad00b26f93b1fcb20b68c622f2"} Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.295609 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cd8c99689-xmlvs" event={"ID":"08b88df2-c726-4578-9f31-ccf6690e58a1","Type":"ContainerStarted","Data":"67090b8143903744524e444b735fcc0af142e75b6700471478ed4bbaee3df2e6"} Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.295621 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cd8c99689-xmlvs" event={"ID":"08b88df2-c726-4578-9f31-ccf6690e58a1","Type":"ContainerStarted","Data":"c6b88266f3769048b7519299b7481b28766d98f699980ef9d5a33673ff972c86"} Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.296541 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.299324 4809 generic.go:334] "Generic (PLEG): container finished" podID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerID="c35995a52f9855a7ef88701b2d863ea4e611683bbda69f5579ae01dd654124bc" exitCode=0 Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.299384 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" event={"ID":"f87668ca-d015-4ac3-b074-a49b7fab8991","Type":"ContainerDied","Data":"c35995a52f9855a7ef88701b2d863ea4e611683bbda69f5579ae01dd654124bc"} Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.299413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" event={"ID":"f87668ca-d015-4ac3-b074-a49b7fab8991","Type":"ContainerStarted","Data":"5d4c76f1016d062d20afb8d8fd9876c39329621ed2b63a5110257552dc64afdc"} Dec 05 12:40:05 crc kubenswrapper[4809]: I1205 12:40:05.321838 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cd8c99689-xmlvs" podStartSLOduration=2.321814342 podStartE2EDuration="2.321814342s" podCreationTimestamp="2025-12-05 12:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:40:05.315075472 +0000 UTC m=+5500.706052030" watchObservedRunningTime="2025-12-05 12:40:05.321814342 +0000 UTC m=+5500.712790900" Dec 05 12:40:06 crc kubenswrapper[4809]: I1205 12:40:06.312114 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" event={"ID":"f87668ca-d015-4ac3-b074-a49b7fab8991","Type":"ContainerStarted","Data":"0cd26242653c78c8a97098fa6b28c64b5ba6eabd8a0a124f658c653f324c804d"} Dec 05 12:40:06 crc kubenswrapper[4809]: I1205 12:40:06.312371 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:06 crc kubenswrapper[4809]: I1205 12:40:06.335583 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" podStartSLOduration=3.335557347 podStartE2EDuration="3.335557347s" podCreationTimestamp="2025-12-05 12:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:40:06.329709271 +0000 UTC m=+5501.720685829" watchObservedRunningTime="2025-12-05 12:40:06.335557347 +0000 UTC m=+5501.726533905" Dec 05 12:40:13 crc kubenswrapper[4809]: I1205 12:40:13.813888 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:40:13 crc kubenswrapper[4809]: I1205 12:40:13.873661 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697b7cf9cc-qwr99"] Dec 05 12:40:13 crc kubenswrapper[4809]: I1205 12:40:13.873914 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" podUID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerName="dnsmasq-dns" containerID="cri-o://f336e1d363bde633a5587d6a97ae454e2a99c824f9f765c91e0f445d4872b5f8" gracePeriod=10 Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.046954 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.047005 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.047047 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.047698 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d29fa2887bb281af96ca44548b0644914998505ad459688152d496f4d53f7af"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.047745 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://0d29fa2887bb281af96ca44548b0644914998505ad459688152d496f4d53f7af" gracePeriod=600 Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.386309 4809 generic.go:334] "Generic (PLEG): container finished" podID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerID="f336e1d363bde633a5587d6a97ae454e2a99c824f9f765c91e0f445d4872b5f8" exitCode=0 Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.386605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" event={"ID":"4fbf0c52-301f-4d29-a333-e8dc4a25448c","Type":"ContainerDied","Data":"f336e1d363bde633a5587d6a97ae454e2a99c824f9f765c91e0f445d4872b5f8"} Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.386645 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" event={"ID":"4fbf0c52-301f-4d29-a333-e8dc4a25448c","Type":"ContainerDied","Data":"ef214ccc672b4bea54af3ec44475e9a10d193c081f8f9cf2dc41cbf871303b8c"} Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.386658 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef214ccc672b4bea54af3ec44475e9a10d193c081f8f9cf2dc41cbf871303b8c" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.386788 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.390308 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="0d29fa2887bb281af96ca44548b0644914998505ad459688152d496f4d53f7af" exitCode=0 Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.390335 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"0d29fa2887bb281af96ca44548b0644914998505ad459688152d496f4d53f7af"} Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.390381 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31"} Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.390400 4809 scope.go:117] "RemoveContainer" containerID="1be431792d028e67378fde9720bd1eb5ce97b04c18a729baebc6f36f4231fd02" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.401469 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-nb\") pod \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.401512 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-sb\") pod \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.401585 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnx2q\" (UniqueName: \"kubernetes.io/projected/4fbf0c52-301f-4d29-a333-e8dc4a25448c-kube-api-access-bnx2q\") pod \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.401693 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-dns-svc\") pod \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.401719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config\") pod \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.412023 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fbf0c52-301f-4d29-a333-e8dc4a25448c-kube-api-access-bnx2q" (OuterVolumeSpecName: "kube-api-access-bnx2q") pod "4fbf0c52-301f-4d29-a333-e8dc4a25448c" (UID: "4fbf0c52-301f-4d29-a333-e8dc4a25448c"). InnerVolumeSpecName "kube-api-access-bnx2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.469333 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4fbf0c52-301f-4d29-a333-e8dc4a25448c" (UID: "4fbf0c52-301f-4d29-a333-e8dc4a25448c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.470056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4fbf0c52-301f-4d29-a333-e8dc4a25448c" (UID: "4fbf0c52-301f-4d29-a333-e8dc4a25448c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:40:14 crc kubenswrapper[4809]: E1205 12:40:14.478891 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config podName:4fbf0c52-301f-4d29-a333-e8dc4a25448c nodeName:}" failed. No retries permitted until 2025-12-05 12:40:14.978859197 +0000 UTC m=+5510.369835765 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config") pod "4fbf0c52-301f-4d29-a333-e8dc4a25448c" (UID: "4fbf0c52-301f-4d29-a333-e8dc4a25448c") : error deleting /var/lib/kubelet/pods/4fbf0c52-301f-4d29-a333-e8dc4a25448c/volume-subpaths: remove /var/lib/kubelet/pods/4fbf0c52-301f-4d29-a333-e8dc4a25448c/volume-subpaths: no such file or directory Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.479084 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4fbf0c52-301f-4d29-a333-e8dc4a25448c" (UID: "4fbf0c52-301f-4d29-a333-e8dc4a25448c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.504294 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnx2q\" (UniqueName: \"kubernetes.io/projected/4fbf0c52-301f-4d29-a333-e8dc4a25448c-kube-api-access-bnx2q\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.504330 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.504342 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:14 crc kubenswrapper[4809]: I1205 12:40:14.504351 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:15 crc kubenswrapper[4809]: I1205 12:40:15.018058 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config\") pod \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\" (UID: \"4fbf0c52-301f-4d29-a333-e8dc4a25448c\") " Dec 05 12:40:15 crc kubenswrapper[4809]: I1205 12:40:15.019690 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config" (OuterVolumeSpecName: "config") pod "4fbf0c52-301f-4d29-a333-e8dc4a25448c" (UID: "4fbf0c52-301f-4d29-a333-e8dc4a25448c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:40:15 crc kubenswrapper[4809]: I1205 12:40:15.120997 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbf0c52-301f-4d29-a333-e8dc4a25448c-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:15 crc kubenswrapper[4809]: I1205 12:40:15.400007 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b7cf9cc-qwr99" Dec 05 12:40:15 crc kubenswrapper[4809]: I1205 12:40:15.436029 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697b7cf9cc-qwr99"] Dec 05 12:40:15 crc kubenswrapper[4809]: I1205 12:40:15.445031 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-697b7cf9cc-qwr99"] Dec 05 12:40:16 crc kubenswrapper[4809]: I1205 12:40:16.883518 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" path="/var/lib/kubelet/pods/4fbf0c52-301f-4d29-a333-e8dc4a25448c/volumes" Dec 05 12:40:33 crc kubenswrapper[4809]: I1205 12:40:33.946597 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cd8c99689-xmlvs" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.504725 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-cpwzd"] Dec 05 12:40:40 crc kubenswrapper[4809]: E1205 12:40:40.505483 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerName="dnsmasq-dns" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.505496 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerName="dnsmasq-dns" Dec 05 12:40:40 crc kubenswrapper[4809]: E1205 12:40:40.505531 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerName="init" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.505537 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerName="init" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.505746 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fbf0c52-301f-4d29-a333-e8dc4a25448c" containerName="dnsmasq-dns" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.506700 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.518818 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cpwzd"] Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.557508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdfdb\" (UniqueName: \"kubernetes.io/projected/7ad6aadd-94ef-435a-a013-d04f6fde42d8-kube-api-access-kdfdb\") pod \"glance-db-create-cpwzd\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.557683 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad6aadd-94ef-435a-a013-d04f6fde42d8-operator-scripts\") pod \"glance-db-create-cpwzd\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.606969 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-85d5-account-create-update-x9nkx"] Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.608072 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.616435 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.618521 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-85d5-account-create-update-x9nkx"] Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.659063 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdfdb\" (UniqueName: \"kubernetes.io/projected/7ad6aadd-94ef-435a-a013-d04f6fde42d8-kube-api-access-kdfdb\") pod \"glance-db-create-cpwzd\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.659446 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad6aadd-94ef-435a-a013-d04f6fde42d8-operator-scripts\") pod \"glance-db-create-cpwzd\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.660277 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad6aadd-94ef-435a-a013-d04f6fde42d8-operator-scripts\") pod \"glance-db-create-cpwzd\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.680587 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdfdb\" (UniqueName: \"kubernetes.io/projected/7ad6aadd-94ef-435a-a013-d04f6fde42d8-kube-api-access-kdfdb\") pod \"glance-db-create-cpwzd\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.761458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btsww\" (UniqueName: \"kubernetes.io/projected/b101ef7e-9934-4bb5-90d5-211a8a42128d-kube-api-access-btsww\") pod \"glance-85d5-account-create-update-x9nkx\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.761517 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b101ef7e-9934-4bb5-90d5-211a8a42128d-operator-scripts\") pod \"glance-85d5-account-create-update-x9nkx\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.829096 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.862834 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btsww\" (UniqueName: \"kubernetes.io/projected/b101ef7e-9934-4bb5-90d5-211a8a42128d-kube-api-access-btsww\") pod \"glance-85d5-account-create-update-x9nkx\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.862905 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b101ef7e-9934-4bb5-90d5-211a8a42128d-operator-scripts\") pod \"glance-85d5-account-create-update-x9nkx\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.863609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b101ef7e-9934-4bb5-90d5-211a8a42128d-operator-scripts\") pod \"glance-85d5-account-create-update-x9nkx\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.883695 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btsww\" (UniqueName: \"kubernetes.io/projected/b101ef7e-9934-4bb5-90d5-211a8a42128d-kube-api-access-btsww\") pod \"glance-85d5-account-create-update-x9nkx\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:40 crc kubenswrapper[4809]: I1205 12:40:40.963809 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.287711 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cpwzd"] Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.439951 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-85d5-account-create-update-x9nkx"] Dec 05 12:40:41 crc kubenswrapper[4809]: W1205 12:40:41.445972 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb101ef7e_9934_4bb5_90d5_211a8a42128d.slice/crio-40878677f9b81b123d1baa73509d9d87893f5b0f703a604957b892e3cc627727 WatchSource:0}: Error finding container 40878677f9b81b123d1baa73509d9d87893f5b0f703a604957b892e3cc627727: Status 404 returned error can't find the container with id 40878677f9b81b123d1baa73509d9d87893f5b0f703a604957b892e3cc627727 Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.727887 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-85d5-account-create-update-x9nkx" event={"ID":"b101ef7e-9934-4bb5-90d5-211a8a42128d","Type":"ContainerStarted","Data":"92b430e4e8024df5ac6b8285abe1bae3e637daeb0d04f4cc6058496cf0999da9"} Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.727936 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-85d5-account-create-update-x9nkx" event={"ID":"b101ef7e-9934-4bb5-90d5-211a8a42128d","Type":"ContainerStarted","Data":"40878677f9b81b123d1baa73509d9d87893f5b0f703a604957b892e3cc627727"} Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.751374 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-85d5-account-create-update-x9nkx" podStartSLOduration=1.751355115 podStartE2EDuration="1.751355115s" podCreationTimestamp="2025-12-05 12:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:40:41.748084508 +0000 UTC m=+5537.139061066" watchObservedRunningTime="2025-12-05 12:40:41.751355115 +0000 UTC m=+5537.142331673" Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.752079 4809 generic.go:334] "Generic (PLEG): container finished" podID="7ad6aadd-94ef-435a-a013-d04f6fde42d8" containerID="aa34681b2e890c1ddab1802193a81d2104d2513559aedcb9cb9471041a924ee0" exitCode=0 Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.752128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cpwzd" event={"ID":"7ad6aadd-94ef-435a-a013-d04f6fde42d8","Type":"ContainerDied","Data":"aa34681b2e890c1ddab1802193a81d2104d2513559aedcb9cb9471041a924ee0"} Dec 05 12:40:41 crc kubenswrapper[4809]: I1205 12:40:41.752155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cpwzd" event={"ID":"7ad6aadd-94ef-435a-a013-d04f6fde42d8","Type":"ContainerStarted","Data":"a2e254f62552f911d1bf6322ad240b5e24a289a4d8a5d5ed1428524dba6be6e8"} Dec 05 12:40:42 crc kubenswrapper[4809]: I1205 12:40:42.764619 4809 generic.go:334] "Generic (PLEG): container finished" podID="b101ef7e-9934-4bb5-90d5-211a8a42128d" containerID="92b430e4e8024df5ac6b8285abe1bae3e637daeb0d04f4cc6058496cf0999da9" exitCode=0 Dec 05 12:40:42 crc kubenswrapper[4809]: I1205 12:40:42.765582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-85d5-account-create-update-x9nkx" event={"ID":"b101ef7e-9934-4bb5-90d5-211a8a42128d","Type":"ContainerDied","Data":"92b430e4e8024df5ac6b8285abe1bae3e637daeb0d04f4cc6058496cf0999da9"} Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.139063 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.224114 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdfdb\" (UniqueName: \"kubernetes.io/projected/7ad6aadd-94ef-435a-a013-d04f6fde42d8-kube-api-access-kdfdb\") pod \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.224250 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad6aadd-94ef-435a-a013-d04f6fde42d8-operator-scripts\") pod \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\" (UID: \"7ad6aadd-94ef-435a-a013-d04f6fde42d8\") " Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.225206 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ad6aadd-94ef-435a-a013-d04f6fde42d8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ad6aadd-94ef-435a-a013-d04f6fde42d8" (UID: "7ad6aadd-94ef-435a-a013-d04f6fde42d8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.230481 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad6aadd-94ef-435a-a013-d04f6fde42d8-kube-api-access-kdfdb" (OuterVolumeSpecName: "kube-api-access-kdfdb") pod "7ad6aadd-94ef-435a-a013-d04f6fde42d8" (UID: "7ad6aadd-94ef-435a-a013-d04f6fde42d8"). InnerVolumeSpecName "kube-api-access-kdfdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.326932 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdfdb\" (UniqueName: \"kubernetes.io/projected/7ad6aadd-94ef-435a-a013-d04f6fde42d8-kube-api-access-kdfdb\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.326990 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad6aadd-94ef-435a-a013-d04f6fde42d8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.775160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cpwzd" event={"ID":"7ad6aadd-94ef-435a-a013-d04f6fde42d8","Type":"ContainerDied","Data":"a2e254f62552f911d1bf6322ad240b5e24a289a4d8a5d5ed1428524dba6be6e8"} Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.775207 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2e254f62552f911d1bf6322ad240b5e24a289a4d8a5d5ed1428524dba6be6e8" Dec 05 12:40:43 crc kubenswrapper[4809]: I1205 12:40:43.775247 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cpwzd" Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.101735 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.137276 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btsww\" (UniqueName: \"kubernetes.io/projected/b101ef7e-9934-4bb5-90d5-211a8a42128d-kube-api-access-btsww\") pod \"b101ef7e-9934-4bb5-90d5-211a8a42128d\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.137408 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b101ef7e-9934-4bb5-90d5-211a8a42128d-operator-scripts\") pod \"b101ef7e-9934-4bb5-90d5-211a8a42128d\" (UID: \"b101ef7e-9934-4bb5-90d5-211a8a42128d\") " Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.138267 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b101ef7e-9934-4bb5-90d5-211a8a42128d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b101ef7e-9934-4bb5-90d5-211a8a42128d" (UID: "b101ef7e-9934-4bb5-90d5-211a8a42128d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.144878 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b101ef7e-9934-4bb5-90d5-211a8a42128d-kube-api-access-btsww" (OuterVolumeSpecName: "kube-api-access-btsww") pod "b101ef7e-9934-4bb5-90d5-211a8a42128d" (UID: "b101ef7e-9934-4bb5-90d5-211a8a42128d"). InnerVolumeSpecName "kube-api-access-btsww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.239249 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b101ef7e-9934-4bb5-90d5-211a8a42128d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.239546 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btsww\" (UniqueName: \"kubernetes.io/projected/b101ef7e-9934-4bb5-90d5-211a8a42128d-kube-api-access-btsww\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.785739 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-85d5-account-create-update-x9nkx" event={"ID":"b101ef7e-9934-4bb5-90d5-211a8a42128d","Type":"ContainerDied","Data":"40878677f9b81b123d1baa73509d9d87893f5b0f703a604957b892e3cc627727"} Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.785780 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40878677f9b81b123d1baa73509d9d87893f5b0f703a604957b892e3cc627727" Dec 05 12:40:44 crc kubenswrapper[4809]: I1205 12:40:44.785853 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-85d5-account-create-update-x9nkx" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.848946 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-n6m5r"] Dec 05 12:40:45 crc kubenswrapper[4809]: E1205 12:40:45.849318 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad6aadd-94ef-435a-a013-d04f6fde42d8" containerName="mariadb-database-create" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.849332 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad6aadd-94ef-435a-a013-d04f6fde42d8" containerName="mariadb-database-create" Dec 05 12:40:45 crc kubenswrapper[4809]: E1205 12:40:45.849346 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b101ef7e-9934-4bb5-90d5-211a8a42128d" containerName="mariadb-account-create-update" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.849352 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b101ef7e-9934-4bb5-90d5-211a8a42128d" containerName="mariadb-account-create-update" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.849518 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b101ef7e-9934-4bb5-90d5-211a8a42128d" containerName="mariadb-account-create-update" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.849542 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad6aadd-94ef-435a-a013-d04f6fde42d8" containerName="mariadb-database-create" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.850242 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.852846 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.853076 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-nhzj2" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.860561 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-n6m5r"] Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.978015 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-db-sync-config-data\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.978416 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-config-data\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.978454 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-combined-ca-bundle\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:45 crc kubenswrapper[4809]: I1205 12:40:45.978552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfp4n\" (UniqueName: \"kubernetes.io/projected/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-kube-api-access-pfp4n\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.079945 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-combined-ca-bundle\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.080215 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfp4n\" (UniqueName: \"kubernetes.io/projected/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-kube-api-access-pfp4n\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.080323 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-db-sync-config-data\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.080478 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-config-data\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.086366 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-config-data\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.086865 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-db-sync-config-data\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.087691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-combined-ca-bundle\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.106753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfp4n\" (UniqueName: \"kubernetes.io/projected/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-kube-api-access-pfp4n\") pod \"glance-db-sync-n6m5r\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.181710 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.718510 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-n6m5r"] Dec 05 12:40:46 crc kubenswrapper[4809]: I1205 12:40:46.805770 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n6m5r" event={"ID":"b944ba07-ed9e-4304-bcb7-8c54e80e27d9","Type":"ContainerStarted","Data":"23d4101cc1e396dad2a1c5a4b1fbce288e89d47e04b6a53ddc031cba15eb9baa"} Dec 05 12:40:47 crc kubenswrapper[4809]: I1205 12:40:47.813287 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n6m5r" event={"ID":"b944ba07-ed9e-4304-bcb7-8c54e80e27d9","Type":"ContainerStarted","Data":"c2b6990d10915b0b0956cf8b9636b9187b3d650df6998b93c6e7e54222a7d4dc"} Dec 05 12:40:47 crc kubenswrapper[4809]: I1205 12:40:47.843148 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-n6m5r" podStartSLOduration=2.843128872 podStartE2EDuration="2.843128872s" podCreationTimestamp="2025-12-05 12:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:40:47.836649728 +0000 UTC m=+5543.227626286" watchObservedRunningTime="2025-12-05 12:40:47.843128872 +0000 UTC m=+5543.234105430" Dec 05 12:40:50 crc kubenswrapper[4809]: I1205 12:40:50.847124 4809 generic.go:334] "Generic (PLEG): container finished" podID="b944ba07-ed9e-4304-bcb7-8c54e80e27d9" containerID="c2b6990d10915b0b0956cf8b9636b9187b3d650df6998b93c6e7e54222a7d4dc" exitCode=0 Dec 05 12:40:50 crc kubenswrapper[4809]: I1205 12:40:50.847242 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n6m5r" event={"ID":"b944ba07-ed9e-4304-bcb7-8c54e80e27d9","Type":"ContainerDied","Data":"c2b6990d10915b0b0956cf8b9636b9187b3d650df6998b93c6e7e54222a7d4dc"} Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.248758 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.385990 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-config-data\") pod \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.386094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-combined-ca-bundle\") pod \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.386147 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-db-sync-config-data\") pod \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.386186 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfp4n\" (UniqueName: \"kubernetes.io/projected/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-kube-api-access-pfp4n\") pod \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\" (UID: \"b944ba07-ed9e-4304-bcb7-8c54e80e27d9\") " Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.390968 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b944ba07-ed9e-4304-bcb7-8c54e80e27d9" (UID: "b944ba07-ed9e-4304-bcb7-8c54e80e27d9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.390991 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-kube-api-access-pfp4n" (OuterVolumeSpecName: "kube-api-access-pfp4n") pod "b944ba07-ed9e-4304-bcb7-8c54e80e27d9" (UID: "b944ba07-ed9e-4304-bcb7-8c54e80e27d9"). InnerVolumeSpecName "kube-api-access-pfp4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.410837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b944ba07-ed9e-4304-bcb7-8c54e80e27d9" (UID: "b944ba07-ed9e-4304-bcb7-8c54e80e27d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.441859 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-config-data" (OuterVolumeSpecName: "config-data") pod "b944ba07-ed9e-4304-bcb7-8c54e80e27d9" (UID: "b944ba07-ed9e-4304-bcb7-8c54e80e27d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.489651 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.489696 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.489727 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.489739 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfp4n\" (UniqueName: \"kubernetes.io/projected/b944ba07-ed9e-4304-bcb7-8c54e80e27d9-kube-api-access-pfp4n\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.866268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n6m5r" event={"ID":"b944ba07-ed9e-4304-bcb7-8c54e80e27d9","Type":"ContainerDied","Data":"23d4101cc1e396dad2a1c5a4b1fbce288e89d47e04b6a53ddc031cba15eb9baa"} Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.866301 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23d4101cc1e396dad2a1c5a4b1fbce288e89d47e04b6a53ddc031cba15eb9baa" Dec 05 12:40:52 crc kubenswrapper[4809]: I1205 12:40:52.866355 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n6m5r" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.238654 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6787b8b8df-5q5ms"] Dec 05 12:40:53 crc kubenswrapper[4809]: E1205 12:40:53.239410 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b944ba07-ed9e-4304-bcb7-8c54e80e27d9" containerName="glance-db-sync" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.239428 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b944ba07-ed9e-4304-bcb7-8c54e80e27d9" containerName="glance-db-sync" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.239723 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b944ba07-ed9e-4304-bcb7-8c54e80e27d9" containerName="glance-db-sync" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.241487 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.254112 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.264718 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.271690 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-nhzj2" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.272170 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.272301 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.272439 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.279915 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6787b8b8df-5q5ms"] Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.305772 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.307750 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-dns-svc\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.307832 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-nb\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.307873 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8pmk\" (UniqueName: \"kubernetes.io/projected/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-kube-api-access-g8pmk\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.307937 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-config\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.329951 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-sb\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.414652 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.416570 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.420404 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432355 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-config-data\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432429 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-dns-svc\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-nb\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432528 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5whc\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-kube-api-access-j5whc\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432568 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8pmk\" (UniqueName: \"kubernetes.io/projected/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-kube-api-access-g8pmk\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432661 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-ceph\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432714 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-logs\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432741 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-scripts\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-config\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.432809 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-sb\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.433252 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.433842 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-sb\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.434325 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-nb\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.434452 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-config\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.437341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-dns-svc\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.463677 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8pmk\" (UniqueName: \"kubernetes.io/projected/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-kube-api-access-g8pmk\") pod \"dnsmasq-dns-6787b8b8df-5q5ms\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534294 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc4xf\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-kube-api-access-dc4xf\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-config-data\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534439 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534458 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5whc\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-kube-api-access-j5whc\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534474 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-logs\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534507 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534542 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534568 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-ceph\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534622 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-logs\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534660 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-scripts\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.534687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.535916 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-logs\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.536436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.540315 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-ceph\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.541780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.542308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-scripts\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.543910 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-config-data\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.555262 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5whc\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-kube-api-access-j5whc\") pod \"glance-default-external-api-0\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.565245 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.601322 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.636069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.636496 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc4xf\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-kube-api-access-dc4xf\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.636609 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.636648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.636667 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-logs\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.636691 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.636742 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.640058 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-logs\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.642103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.644266 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.644313 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.645898 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.646525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.659695 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc4xf\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-kube-api-access-dc4xf\") pod \"glance-default-internal-api-0\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:53 crc kubenswrapper[4809]: I1205 12:40:53.756881 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.081042 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6787b8b8df-5q5ms"] Dec 05 12:40:54 crc kubenswrapper[4809]: W1205 12:40:54.085102 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f7a4cca_bd30_4a4b_a841_d37c3e8cc4f6.slice/crio-72db3dbe88e35211b49e63f580f2ee04e8740b363173aca51268f882155b703a WatchSource:0}: Error finding container 72db3dbe88e35211b49e63f580f2ee04e8740b363173aca51268f882155b703a: Status 404 returned error can't find the container with id 72db3dbe88e35211b49e63f580f2ee04e8740b363173aca51268f882155b703a Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.375717 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:54 crc kubenswrapper[4809]: W1205 12:40:54.410779 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90495f0c_98ae_4e7c_a180_07bd8f0fa716.slice/crio-52bb278fc1c49fcda0ec7936471d87632b95c062bde206e089155fbe0e7bc9c0 WatchSource:0}: Error finding container 52bb278fc1c49fcda0ec7936471d87632b95c062bde206e089155fbe0e7bc9c0: Status 404 returned error can't find the container with id 52bb278fc1c49fcda0ec7936471d87632b95c062bde206e089155fbe0e7bc9c0 Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.427321 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.893746 4809 generic.go:334] "Generic (PLEG): container finished" podID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerID="f432078dbf40f8bd74f484a7d063b1566e923ba530c0c3cd719112d6ba379fd1" exitCode=0 Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.901283 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90495f0c-98ae-4e7c-a180-07bd8f0fa716","Type":"ContainerStarted","Data":"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d"} Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.901334 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90495f0c-98ae-4e7c-a180-07bd8f0fa716","Type":"ContainerStarted","Data":"52bb278fc1c49fcda0ec7936471d87632b95c062bde206e089155fbe0e7bc9c0"} Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.901347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" event={"ID":"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6","Type":"ContainerDied","Data":"f432078dbf40f8bd74f484a7d063b1566e923ba530c0c3cd719112d6ba379fd1"} Dec 05 12:40:54 crc kubenswrapper[4809]: I1205 12:40:54.901365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" event={"ID":"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6","Type":"ContainerStarted","Data":"72db3dbe88e35211b49e63f580f2ee04e8740b363173aca51268f882155b703a"} Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.091141 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:55 crc kubenswrapper[4809]: W1205 12:40:55.096323 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87adb1b0_a1b2_49a1_8e3f_3c0e6461a37d.slice/crio-70fd52814b751394b8ea9edc30006d31993b9be4fa6edd35a53f7ca1c3dbe74c WatchSource:0}: Error finding container 70fd52814b751394b8ea9edc30006d31993b9be4fa6edd35a53f7ca1c3dbe74c: Status 404 returned error can't find the container with id 70fd52814b751394b8ea9edc30006d31993b9be4fa6edd35a53f7ca1c3dbe74c Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.906728 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" event={"ID":"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6","Type":"ContainerStarted","Data":"61469d6fc98bdbbf5f3b16a26a2cb939a1844e9ec99bfb61bb3847bc9db963f3"} Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.907100 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.913179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90495f0c-98ae-4e7c-a180-07bd8f0fa716","Type":"ContainerStarted","Data":"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954"} Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.913376 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-log" containerID="cri-o://403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d" gracePeriod=30 Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.913686 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-httpd" containerID="cri-o://b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954" gracePeriod=30 Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.920665 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d","Type":"ContainerStarted","Data":"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498"} Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.920723 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d","Type":"ContainerStarted","Data":"70fd52814b751394b8ea9edc30006d31993b9be4fa6edd35a53f7ca1c3dbe74c"} Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.933265 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" podStartSLOduration=2.9332440760000003 podStartE2EDuration="2.933244076s" podCreationTimestamp="2025-12-05 12:40:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:40:55.92514371 +0000 UTC m=+5551.316120288" watchObservedRunningTime="2025-12-05 12:40:55.933244076 +0000 UTC m=+5551.324220634" Dec 05 12:40:55 crc kubenswrapper[4809]: I1205 12:40:55.951027 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.951006122 podStartE2EDuration="2.951006122s" podCreationTimestamp="2025-12-05 12:40:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:40:55.945739191 +0000 UTC m=+5551.336715769" watchObservedRunningTime="2025-12-05 12:40:55.951006122 +0000 UTC m=+5551.341982680" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.183779 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.583146 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.698593 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-scripts\") pod \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.698764 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-combined-ca-bundle\") pod \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.698795 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-logs\") pod \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.698817 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-httpd-run\") pod \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.698878 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5whc\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-kube-api-access-j5whc\") pod \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.698936 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-config-data\") pod \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.698961 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-ceph\") pod \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\" (UID: \"90495f0c-98ae-4e7c-a180-07bd8f0fa716\") " Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.700571 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-logs" (OuterVolumeSpecName: "logs") pod "90495f0c-98ae-4e7c-a180-07bd8f0fa716" (UID: "90495f0c-98ae-4e7c-a180-07bd8f0fa716"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.700864 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "90495f0c-98ae-4e7c-a180-07bd8f0fa716" (UID: "90495f0c-98ae-4e7c-a180-07bd8f0fa716"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.707714 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-kube-api-access-j5whc" (OuterVolumeSpecName: "kube-api-access-j5whc") pod "90495f0c-98ae-4e7c-a180-07bd8f0fa716" (UID: "90495f0c-98ae-4e7c-a180-07bd8f0fa716"). InnerVolumeSpecName "kube-api-access-j5whc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.707751 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-scripts" (OuterVolumeSpecName: "scripts") pod "90495f0c-98ae-4e7c-a180-07bd8f0fa716" (UID: "90495f0c-98ae-4e7c-a180-07bd8f0fa716"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.707819 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-ceph" (OuterVolumeSpecName: "ceph") pod "90495f0c-98ae-4e7c-a180-07bd8f0fa716" (UID: "90495f0c-98ae-4e7c-a180-07bd8f0fa716"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.739773 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90495f0c-98ae-4e7c-a180-07bd8f0fa716" (UID: "90495f0c-98ae-4e7c-a180-07bd8f0fa716"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.754919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-config-data" (OuterVolumeSpecName: "config-data") pod "90495f0c-98ae-4e7c-a180-07bd8f0fa716" (UID: "90495f0c-98ae-4e7c-a180-07bd8f0fa716"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.800801 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5whc\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-kube-api-access-j5whc\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.800992 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.801069 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/90495f0c-98ae-4e7c-a180-07bd8f0fa716-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.801394 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.801480 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90495f0c-98ae-4e7c-a180-07bd8f0fa716-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.801559 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.801672 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/90495f0c-98ae-4e7c-a180-07bd8f0fa716-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.931028 4809 generic.go:334] "Generic (PLEG): container finished" podID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerID="b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954" exitCode=0 Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.931077 4809 generic.go:334] "Generic (PLEG): container finished" podID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerID="403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d" exitCode=143 Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.931085 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.931147 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90495f0c-98ae-4e7c-a180-07bd8f0fa716","Type":"ContainerDied","Data":"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954"} Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.931182 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90495f0c-98ae-4e7c-a180-07bd8f0fa716","Type":"ContainerDied","Data":"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d"} Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.931195 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"90495f0c-98ae-4e7c-a180-07bd8f0fa716","Type":"ContainerDied","Data":"52bb278fc1c49fcda0ec7936471d87632b95c062bde206e089155fbe0e7bc9c0"} Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.931229 4809 scope.go:117] "RemoveContainer" containerID="b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.936986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d","Type":"ContainerStarted","Data":"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17"} Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.961458 4809 scope.go:117] "RemoveContainer" containerID="403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.971800 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.980047 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.987017 4809 scope.go:117] "RemoveContainer" containerID="b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954" Dec 05 12:40:56 crc kubenswrapper[4809]: E1205 12:40:56.989300 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954\": container with ID starting with b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954 not found: ID does not exist" containerID="b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.989347 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954"} err="failed to get container status \"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954\": rpc error: code = NotFound desc = could not find container \"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954\": container with ID starting with b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954 not found: ID does not exist" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.989378 4809 scope.go:117] "RemoveContainer" containerID="403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d" Dec 05 12:40:56 crc kubenswrapper[4809]: E1205 12:40:56.989676 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d\": container with ID starting with 403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d not found: ID does not exist" containerID="403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.989721 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d"} err="failed to get container status \"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d\": rpc error: code = NotFound desc = could not find container \"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d\": container with ID starting with 403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d not found: ID does not exist" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.989741 4809 scope.go:117] "RemoveContainer" containerID="b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.991840 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954"} err="failed to get container status \"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954\": rpc error: code = NotFound desc = could not find container \"b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954\": container with ID starting with b48debd20e2955bfef377d51d8e88e81a0073871cad9d52648f8d5b74c66b954 not found: ID does not exist" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.991872 4809 scope.go:117] "RemoveContainer" containerID="403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d" Dec 05 12:40:56 crc kubenswrapper[4809]: I1205 12:40:56.992695 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d"} err="failed to get container status \"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d\": rpc error: code = NotFound desc = could not find container \"403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d\": container with ID starting with 403de0dc07a1c7db29e25dd833ac6523e430a2799f8425c1c1be97a1e454ac7d not found: ID does not exist" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.007822 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.007801129 podStartE2EDuration="4.007801129s" podCreationTimestamp="2025-12-05 12:40:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:40:56.987300261 +0000 UTC m=+5552.378276819" watchObservedRunningTime="2025-12-05 12:40:57.007801129 +0000 UTC m=+5552.398777677" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.016899 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:57 crc kubenswrapper[4809]: E1205 12:40:57.017604 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-log" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.017622 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-log" Dec 05 12:40:57 crc kubenswrapper[4809]: E1205 12:40:57.017659 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-httpd" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.017666 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-httpd" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.017839 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-httpd" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.017863 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" containerName="glance-log" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.018771 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.023564 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.041778 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.108291 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-ceph\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.108357 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.108397 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-logs\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.108432 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.108484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t47hv\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-kube-api-access-t47hv\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.108529 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-scripts\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.108571 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-config-data\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.210479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-ceph\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.210536 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.210575 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-logs\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.210605 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.210665 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t47hv\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-kube-api-access-t47hv\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.210694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-scripts\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.210727 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-config-data\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.211550 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-logs\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.212099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.214664 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-scripts\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.215576 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-ceph\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.216573 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-config-data\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.217401 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.232921 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t47hv\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-kube-api-access-t47hv\") pod \"glance-default-external-api-0\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.348108 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.918176 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:40:57 crc kubenswrapper[4809]: W1205 12:40:57.919778 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb435d598_7879_4eeb_97fa_91d8df2ec977.slice/crio-5b74067b930ef3e0a28930e6256c29e661bd4876b580eea7a64c79bf6e623cd6 WatchSource:0}: Error finding container 5b74067b930ef3e0a28930e6256c29e661bd4876b580eea7a64c79bf6e623cd6: Status 404 returned error can't find the container with id 5b74067b930ef3e0a28930e6256c29e661bd4876b580eea7a64c79bf6e623cd6 Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.946112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b435d598-7879-4eeb-97fa-91d8df2ec977","Type":"ContainerStarted","Data":"5b74067b930ef3e0a28930e6256c29e661bd4876b580eea7a64c79bf6e623cd6"} Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.948266 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-log" containerID="cri-o://291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498" gracePeriod=30 Dec 05 12:40:57 crc kubenswrapper[4809]: I1205 12:40:57.948335 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-httpd" containerID="cri-o://f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17" gracePeriod=30 Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.471836 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.535578 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-logs\") pod \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.535671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc4xf\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-kube-api-access-dc4xf\") pod \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.535805 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-ceph\") pod \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.535856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-httpd-run\") pod \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.535886 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-combined-ca-bundle\") pod \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.535951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-config-data\") pod \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.536204 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-logs" (OuterVolumeSpecName: "logs") pod "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" (UID: "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.536577 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" (UID: "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.536927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-scripts\") pod \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\" (UID: \"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d\") " Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.537635 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.537742 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.541716 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-kube-api-access-dc4xf" (OuterVolumeSpecName: "kube-api-access-dc4xf") pod "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" (UID: "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d"). InnerVolumeSpecName "kube-api-access-dc4xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.542634 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-scripts" (OuterVolumeSpecName: "scripts") pod "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" (UID: "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.555193 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-ceph" (OuterVolumeSpecName: "ceph") pod "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" (UID: "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.593842 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" (UID: "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.611546 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-config-data" (OuterVolumeSpecName: "config-data") pod "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" (UID: "87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.639259 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.639288 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.639299 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.639308 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc4xf\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-kube-api-access-dc4xf\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.639319 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.884728 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90495f0c-98ae-4e7c-a180-07bd8f0fa716" path="/var/lib/kubelet/pods/90495f0c-98ae-4e7c-a180-07bd8f0fa716/volumes" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.961422 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b435d598-7879-4eeb-97fa-91d8df2ec977","Type":"ContainerStarted","Data":"825928806dff97aafe8db5614b26d9c6e3400f3755111b41a90eaf2c5c096d56"} Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.963140 4809 generic.go:334] "Generic (PLEG): container finished" podID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerID="f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17" exitCode=0 Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.963162 4809 generic.go:334] "Generic (PLEG): container finished" podID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerID="291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498" exitCode=143 Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.963177 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d","Type":"ContainerDied","Data":"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17"} Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.963194 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d","Type":"ContainerDied","Data":"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498"} Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.963203 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d","Type":"ContainerDied","Data":"70fd52814b751394b8ea9edc30006d31993b9be4fa6edd35a53f7ca1c3dbe74c"} Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.963215 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.963220 4809 scope.go:117] "RemoveContainer" containerID="f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17" Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.987314 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:58 crc kubenswrapper[4809]: I1205 12:40:58.996840 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.008112 4809 scope.go:117] "RemoveContainer" containerID="291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.012821 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:59 crc kubenswrapper[4809]: E1205 12:40:59.013903 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-httpd" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.014084 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-httpd" Dec 05 12:40:59 crc kubenswrapper[4809]: E1205 12:40:59.014215 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-log" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.014299 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-log" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.014605 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-log" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.014732 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" containerName="glance-httpd" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.016057 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.018134 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.055070 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.088657 4809 scope.go:117] "RemoveContainer" containerID="f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17" Dec 05 12:40:59 crc kubenswrapper[4809]: E1205 12:40:59.089333 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17\": container with ID starting with f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17 not found: ID does not exist" containerID="f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.089378 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17"} err="failed to get container status \"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17\": rpc error: code = NotFound desc = could not find container \"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17\": container with ID starting with f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17 not found: ID does not exist" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.089409 4809 scope.go:117] "RemoveContainer" containerID="291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498" Dec 05 12:40:59 crc kubenswrapper[4809]: E1205 12:40:59.090047 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498\": container with ID starting with 291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498 not found: ID does not exist" containerID="291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.090075 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498"} err="failed to get container status \"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498\": rpc error: code = NotFound desc = could not find container \"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498\": container with ID starting with 291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498 not found: ID does not exist" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.090089 4809 scope.go:117] "RemoveContainer" containerID="f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.090547 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17"} err="failed to get container status \"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17\": rpc error: code = NotFound desc = could not find container \"f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17\": container with ID starting with f4e8eef139f61e89f5137e128d08d4748245f5d2d25fb6b40cefc3e7ff148c17 not found: ID does not exist" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.090626 4809 scope.go:117] "RemoveContainer" containerID="291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.091221 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498"} err="failed to get container status \"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498\": rpc error: code = NotFound desc = could not find container \"291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498\": container with ID starting with 291f2581ae2117a5b70f1ec32870bd44709a21d23bcd6aa5ee1ce3bc59517498 not found: ID does not exist" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.147457 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.147540 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.147591 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.147615 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-logs\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.147734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s46kl\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-kube-api-access-s46kl\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.147755 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.148601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.249758 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s46kl\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-kube-api-access-s46kl\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.250116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.250174 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.250232 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.250278 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.250305 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.250326 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-logs\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.250942 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.251197 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-logs\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.261728 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.261753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.261869 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.262277 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.269810 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s46kl\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-kube-api-access-s46kl\") pod \"glance-default-internal-api-0\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.353687 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:40:59 crc kubenswrapper[4809]: W1205 12:40:59.906873 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbf81b4e_4c9d_4ee9_8149_fc17884aa78e.slice/crio-9d03496bc2ff9d0d6d8977764b2367ac63eff8fef327619da24bb1fd7b4db617 WatchSource:0}: Error finding container 9d03496bc2ff9d0d6d8977764b2367ac63eff8fef327619da24bb1fd7b4db617: Status 404 returned error can't find the container with id 9d03496bc2ff9d0d6d8977764b2367ac63eff8fef327619da24bb1fd7b4db617 Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.907130 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.976324 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e","Type":"ContainerStarted","Data":"9d03496bc2ff9d0d6d8977764b2367ac63eff8fef327619da24bb1fd7b4db617"} Dec 05 12:40:59 crc kubenswrapper[4809]: I1205 12:40:59.978956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b435d598-7879-4eeb-97fa-91d8df2ec977","Type":"ContainerStarted","Data":"7e9e0d1c4054e24fb71232d57d5d2ed077ca1fcafa3b53fcda48f9701a1b0aa8"} Dec 05 12:41:00 crc kubenswrapper[4809]: I1205 12:41:00.022998 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.022979829 podStartE2EDuration="4.022979829s" podCreationTimestamp="2025-12-05 12:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:41:00.015209751 +0000 UTC m=+5555.406186309" watchObservedRunningTime="2025-12-05 12:41:00.022979829 +0000 UTC m=+5555.413956387" Dec 05 12:41:00 crc kubenswrapper[4809]: I1205 12:41:00.884894 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d" path="/var/lib/kubelet/pods/87adb1b0-a1b2-49a1-8e3f-3c0e6461a37d/volumes" Dec 05 12:41:00 crc kubenswrapper[4809]: I1205 12:41:00.989228 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e","Type":"ContainerStarted","Data":"f9a990752c48371d529e3c5e427ac07606f04f0669e1a6d399824b2077f1a7a8"} Dec 05 12:41:01 crc kubenswrapper[4809]: I1205 12:41:01.999879 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e","Type":"ContainerStarted","Data":"295010b98790218c4d1fe7af3615987df15f65937024f30689173d2a697cb687"} Dec 05 12:41:02 crc kubenswrapper[4809]: I1205 12:41:02.023581 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.023559981 podStartE2EDuration="4.023559981s" podCreationTimestamp="2025-12-05 12:40:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:41:02.022958525 +0000 UTC m=+5557.413935113" watchObservedRunningTime="2025-12-05 12:41:02.023559981 +0000 UTC m=+5557.414536539" Dec 05 12:41:03 crc kubenswrapper[4809]: I1205 12:41:03.566813 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:41:03 crc kubenswrapper[4809]: I1205 12:41:03.645671 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84c6c75d89-m7nqm"] Dec 05 12:41:03 crc kubenswrapper[4809]: I1205 12:41:03.645973 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerName="dnsmasq-dns" containerID="cri-o://0cd26242653c78c8a97098fa6b28c64b5ba6eabd8a0a124f658c653f324c804d" gracePeriod=10 Dec 05 12:41:03 crc kubenswrapper[4809]: I1205 12:41:03.812928 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.38:5353: connect: connection refused" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.021437 4809 generic.go:334] "Generic (PLEG): container finished" podID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerID="0cd26242653c78c8a97098fa6b28c64b5ba6eabd8a0a124f658c653f324c804d" exitCode=0 Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.021486 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" event={"ID":"f87668ca-d015-4ac3-b074-a49b7fab8991","Type":"ContainerDied","Data":"0cd26242653c78c8a97098fa6b28c64b5ba6eabd8a0a124f658c653f324c804d"} Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.021517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" event={"ID":"f87668ca-d015-4ac3-b074-a49b7fab8991","Type":"ContainerDied","Data":"5d4c76f1016d062d20afb8d8fd9876c39329621ed2b63a5110257552dc64afdc"} Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.021531 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d4c76f1016d062d20afb8d8fd9876c39329621ed2b63a5110257552dc64afdc" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.087457 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.147211 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-config\") pod \"f87668ca-d015-4ac3-b074-a49b7fab8991\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.147259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-sb\") pod \"f87668ca-d015-4ac3-b074-a49b7fab8991\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.147338 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-dns-svc\") pod \"f87668ca-d015-4ac3-b074-a49b7fab8991\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.147370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4sb9\" (UniqueName: \"kubernetes.io/projected/f87668ca-d015-4ac3-b074-a49b7fab8991-kube-api-access-n4sb9\") pod \"f87668ca-d015-4ac3-b074-a49b7fab8991\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.147440 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-nb\") pod \"f87668ca-d015-4ac3-b074-a49b7fab8991\" (UID: \"f87668ca-d015-4ac3-b074-a49b7fab8991\") " Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.152247 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f87668ca-d015-4ac3-b074-a49b7fab8991-kube-api-access-n4sb9" (OuterVolumeSpecName: "kube-api-access-n4sb9") pod "f87668ca-d015-4ac3-b074-a49b7fab8991" (UID: "f87668ca-d015-4ac3-b074-a49b7fab8991"). InnerVolumeSpecName "kube-api-access-n4sb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.192113 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f87668ca-d015-4ac3-b074-a49b7fab8991" (UID: "f87668ca-d015-4ac3-b074-a49b7fab8991"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.199086 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f87668ca-d015-4ac3-b074-a49b7fab8991" (UID: "f87668ca-d015-4ac3-b074-a49b7fab8991"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.199289 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f87668ca-d015-4ac3-b074-a49b7fab8991" (UID: "f87668ca-d015-4ac3-b074-a49b7fab8991"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.200355 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-config" (OuterVolumeSpecName: "config") pod "f87668ca-d015-4ac3-b074-a49b7fab8991" (UID: "f87668ca-d015-4ac3-b074-a49b7fab8991"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.250011 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.250059 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.250072 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.250084 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4sb9\" (UniqueName: \"kubernetes.io/projected/f87668ca-d015-4ac3-b074-a49b7fab8991-kube-api-access-n4sb9\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:04 crc kubenswrapper[4809]: I1205 12:41:04.250097 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f87668ca-d015-4ac3-b074-a49b7fab8991-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:05 crc kubenswrapper[4809]: I1205 12:41:05.030084 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84c6c75d89-m7nqm" Dec 05 12:41:05 crc kubenswrapper[4809]: I1205 12:41:05.050588 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84c6c75d89-m7nqm"] Dec 05 12:41:05 crc kubenswrapper[4809]: I1205 12:41:05.057520 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84c6c75d89-m7nqm"] Dec 05 12:41:06 crc kubenswrapper[4809]: I1205 12:41:06.885232 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" path="/var/lib/kubelet/pods/f87668ca-d015-4ac3-b074-a49b7fab8991/volumes" Dec 05 12:41:07 crc kubenswrapper[4809]: I1205 12:41:07.348861 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 12:41:07 crc kubenswrapper[4809]: I1205 12:41:07.349186 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 12:41:07 crc kubenswrapper[4809]: I1205 12:41:07.376437 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 12:41:07 crc kubenswrapper[4809]: I1205 12:41:07.387383 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 12:41:08 crc kubenswrapper[4809]: I1205 12:41:08.071870 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 12:41:08 crc kubenswrapper[4809]: I1205 12:41:08.071935 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 12:41:09 crc kubenswrapper[4809]: I1205 12:41:09.354563 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:09 crc kubenswrapper[4809]: I1205 12:41:09.354916 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:09 crc kubenswrapper[4809]: I1205 12:41:09.400572 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:09 crc kubenswrapper[4809]: I1205 12:41:09.421225 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:10 crc kubenswrapper[4809]: I1205 12:41:10.088433 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:10 crc kubenswrapper[4809]: I1205 12:41:10.088465 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:10 crc kubenswrapper[4809]: I1205 12:41:10.395033 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 12:41:10 crc kubenswrapper[4809]: I1205 12:41:10.395453 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:41:10 crc kubenswrapper[4809]: I1205 12:41:10.398566 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 12:41:12 crc kubenswrapper[4809]: I1205 12:41:12.102857 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:41:12 crc kubenswrapper[4809]: I1205 12:41:12.103371 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:41:12 crc kubenswrapper[4809]: I1205 12:41:12.180086 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:12 crc kubenswrapper[4809]: I1205 12:41:12.280788 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.027465 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wgztn"] Dec 05 12:41:18 crc kubenswrapper[4809]: E1205 12:41:18.028418 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerName="dnsmasq-dns" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.028436 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerName="dnsmasq-dns" Dec 05 12:41:18 crc kubenswrapper[4809]: E1205 12:41:18.028459 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerName="init" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.028467 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerName="init" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.028719 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f87668ca-d015-4ac3-b074-a49b7fab8991" containerName="dnsmasq-dns" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.029485 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.078258 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wgztn"] Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.110233 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d28bm\" (UniqueName: \"kubernetes.io/projected/3103bf31-c6f8-494f-a75e-deb3cea6d915-kube-api-access-d28bm\") pod \"placement-db-create-wgztn\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.110423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3103bf31-c6f8-494f-a75e-deb3cea6d915-operator-scripts\") pod \"placement-db-create-wgztn\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.125549 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6087-account-create-update-pbc2b"] Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.127264 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.133661 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.146556 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6087-account-create-update-pbc2b"] Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.212423 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-operator-scripts\") pod \"placement-6087-account-create-update-pbc2b\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.212815 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3103bf31-c6f8-494f-a75e-deb3cea6d915-operator-scripts\") pod \"placement-db-create-wgztn\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.212991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7zvd\" (UniqueName: \"kubernetes.io/projected/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-kube-api-access-c7zvd\") pod \"placement-6087-account-create-update-pbc2b\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.213167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d28bm\" (UniqueName: \"kubernetes.io/projected/3103bf31-c6f8-494f-a75e-deb3cea6d915-kube-api-access-d28bm\") pod \"placement-db-create-wgztn\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.213756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3103bf31-c6f8-494f-a75e-deb3cea6d915-operator-scripts\") pod \"placement-db-create-wgztn\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.236332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d28bm\" (UniqueName: \"kubernetes.io/projected/3103bf31-c6f8-494f-a75e-deb3cea6d915-kube-api-access-d28bm\") pod \"placement-db-create-wgztn\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.315594 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-operator-scripts\") pod \"placement-6087-account-create-update-pbc2b\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.315715 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7zvd\" (UniqueName: \"kubernetes.io/projected/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-kube-api-access-c7zvd\") pod \"placement-6087-account-create-update-pbc2b\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.316742 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-operator-scripts\") pod \"placement-6087-account-create-update-pbc2b\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.357297 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7zvd\" (UniqueName: \"kubernetes.io/projected/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-kube-api-access-c7zvd\") pod \"placement-6087-account-create-update-pbc2b\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.366139 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wgztn" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.449704 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.933421 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wgztn"] Dec 05 12:41:18 crc kubenswrapper[4809]: W1205 12:41:18.934235 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3103bf31_c6f8_494f_a75e_deb3cea6d915.slice/crio-ac30ae3ad0e24efb73b186c3d596c4b1ff14ca0b001f5df54368a56c1c6b90e4 WatchSource:0}: Error finding container ac30ae3ad0e24efb73b186c3d596c4b1ff14ca0b001f5df54368a56c1c6b90e4: Status 404 returned error can't find the container with id ac30ae3ad0e24efb73b186c3d596c4b1ff14ca0b001f5df54368a56c1c6b90e4 Dec 05 12:41:18 crc kubenswrapper[4809]: I1205 12:41:18.946965 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6087-account-create-update-pbc2b"] Dec 05 12:41:19 crc kubenswrapper[4809]: I1205 12:41:19.164548 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6087-account-create-update-pbc2b" event={"ID":"4b95a840-4da9-4162-bf78-a89b0c0ce3ba","Type":"ContainerStarted","Data":"029e738f53db4fb600d138c70628dbfb950fc25e99085270e213d435231b9298"} Dec 05 12:41:19 crc kubenswrapper[4809]: I1205 12:41:19.164992 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6087-account-create-update-pbc2b" event={"ID":"4b95a840-4da9-4162-bf78-a89b0c0ce3ba","Type":"ContainerStarted","Data":"525f91561f708310e75cb4c467dc41b89ea1fd6a7ffff5a53b87db223d30ea51"} Dec 05 12:41:19 crc kubenswrapper[4809]: I1205 12:41:19.166292 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wgztn" event={"ID":"3103bf31-c6f8-494f-a75e-deb3cea6d915","Type":"ContainerStarted","Data":"6a3c444356bd95b62b53d5c7987521329b1cd06773c88ece1cee34b84341a42e"} Dec 05 12:41:19 crc kubenswrapper[4809]: I1205 12:41:19.166342 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wgztn" event={"ID":"3103bf31-c6f8-494f-a75e-deb3cea6d915","Type":"ContainerStarted","Data":"ac30ae3ad0e24efb73b186c3d596c4b1ff14ca0b001f5df54368a56c1c6b90e4"} Dec 05 12:41:19 crc kubenswrapper[4809]: I1205 12:41:19.189888 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6087-account-create-update-pbc2b" podStartSLOduration=1.189870229 podStartE2EDuration="1.189870229s" podCreationTimestamp="2025-12-05 12:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:41:19.185941604 +0000 UTC m=+5574.576918162" watchObservedRunningTime="2025-12-05 12:41:19.189870229 +0000 UTC m=+5574.580846787" Dec 05 12:41:19 crc kubenswrapper[4809]: I1205 12:41:19.204272 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-wgztn" podStartSLOduration=1.2042503230000001 podStartE2EDuration="1.204250323s" podCreationTimestamp="2025-12-05 12:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:41:19.198287604 +0000 UTC m=+5574.589264152" watchObservedRunningTime="2025-12-05 12:41:19.204250323 +0000 UTC m=+5574.595226881" Dec 05 12:41:20 crc kubenswrapper[4809]: I1205 12:41:20.178688 4809 generic.go:334] "Generic (PLEG): container finished" podID="4b95a840-4da9-4162-bf78-a89b0c0ce3ba" containerID="029e738f53db4fb600d138c70628dbfb950fc25e99085270e213d435231b9298" exitCode=0 Dec 05 12:41:20 crc kubenswrapper[4809]: I1205 12:41:20.178829 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6087-account-create-update-pbc2b" event={"ID":"4b95a840-4da9-4162-bf78-a89b0c0ce3ba","Type":"ContainerDied","Data":"029e738f53db4fb600d138c70628dbfb950fc25e99085270e213d435231b9298"} Dec 05 12:41:20 crc kubenswrapper[4809]: I1205 12:41:20.181028 4809 generic.go:334] "Generic (PLEG): container finished" podID="3103bf31-c6f8-494f-a75e-deb3cea6d915" containerID="6a3c444356bd95b62b53d5c7987521329b1cd06773c88ece1cee34b84341a42e" exitCode=0 Dec 05 12:41:20 crc kubenswrapper[4809]: I1205 12:41:20.181066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wgztn" event={"ID":"3103bf31-c6f8-494f-a75e-deb3cea6d915","Type":"ContainerDied","Data":"6a3c444356bd95b62b53d5c7987521329b1cd06773c88ece1cee34b84341a42e"} Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.640970 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.647666 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wgztn" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.797065 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3103bf31-c6f8-494f-a75e-deb3cea6d915-operator-scripts\") pod \"3103bf31-c6f8-494f-a75e-deb3cea6d915\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.797259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7zvd\" (UniqueName: \"kubernetes.io/projected/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-kube-api-access-c7zvd\") pod \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.797315 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d28bm\" (UniqueName: \"kubernetes.io/projected/3103bf31-c6f8-494f-a75e-deb3cea6d915-kube-api-access-d28bm\") pod \"3103bf31-c6f8-494f-a75e-deb3cea6d915\" (UID: \"3103bf31-c6f8-494f-a75e-deb3cea6d915\") " Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.797362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-operator-scripts\") pod \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\" (UID: \"4b95a840-4da9-4162-bf78-a89b0c0ce3ba\") " Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.798095 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3103bf31-c6f8-494f-a75e-deb3cea6d915-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3103bf31-c6f8-494f-a75e-deb3cea6d915" (UID: "3103bf31-c6f8-494f-a75e-deb3cea6d915"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.798144 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4b95a840-4da9-4162-bf78-a89b0c0ce3ba" (UID: "4b95a840-4da9-4162-bf78-a89b0c0ce3ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.802896 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-kube-api-access-c7zvd" (OuterVolumeSpecName: "kube-api-access-c7zvd") pod "4b95a840-4da9-4162-bf78-a89b0c0ce3ba" (UID: "4b95a840-4da9-4162-bf78-a89b0c0ce3ba"). InnerVolumeSpecName "kube-api-access-c7zvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.803032 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3103bf31-c6f8-494f-a75e-deb3cea6d915-kube-api-access-d28bm" (OuterVolumeSpecName: "kube-api-access-d28bm") pod "3103bf31-c6f8-494f-a75e-deb3cea6d915" (UID: "3103bf31-c6f8-494f-a75e-deb3cea6d915"). InnerVolumeSpecName "kube-api-access-d28bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.899209 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7zvd\" (UniqueName: \"kubernetes.io/projected/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-kube-api-access-c7zvd\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.899505 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d28bm\" (UniqueName: \"kubernetes.io/projected/3103bf31-c6f8-494f-a75e-deb3cea6d915-kube-api-access-d28bm\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.899515 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b95a840-4da9-4162-bf78-a89b0c0ce3ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:21 crc kubenswrapper[4809]: I1205 12:41:21.899524 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3103bf31-c6f8-494f-a75e-deb3cea6d915-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:22 crc kubenswrapper[4809]: I1205 12:41:22.197863 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6087-account-create-update-pbc2b" event={"ID":"4b95a840-4da9-4162-bf78-a89b0c0ce3ba","Type":"ContainerDied","Data":"525f91561f708310e75cb4c467dc41b89ea1fd6a7ffff5a53b87db223d30ea51"} Dec 05 12:41:22 crc kubenswrapper[4809]: I1205 12:41:22.197889 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6087-account-create-update-pbc2b" Dec 05 12:41:22 crc kubenswrapper[4809]: I1205 12:41:22.197914 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="525f91561f708310e75cb4c467dc41b89ea1fd6a7ffff5a53b87db223d30ea51" Dec 05 12:41:22 crc kubenswrapper[4809]: I1205 12:41:22.199365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wgztn" event={"ID":"3103bf31-c6f8-494f-a75e-deb3cea6d915","Type":"ContainerDied","Data":"ac30ae3ad0e24efb73b186c3d596c4b1ff14ca0b001f5df54368a56c1c6b90e4"} Dec 05 12:41:22 crc kubenswrapper[4809]: I1205 12:41:22.199399 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac30ae3ad0e24efb73b186c3d596c4b1ff14ca0b001f5df54368a56c1c6b90e4" Dec 05 12:41:22 crc kubenswrapper[4809]: I1205 12:41:22.199465 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wgztn" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.539729 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f894f5fcc-p99x8"] Dec 05 12:41:23 crc kubenswrapper[4809]: E1205 12:41:23.540132 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b95a840-4da9-4162-bf78-a89b0c0ce3ba" containerName="mariadb-account-create-update" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.540423 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b95a840-4da9-4162-bf78-a89b0c0ce3ba" containerName="mariadb-account-create-update" Dec 05 12:41:23 crc kubenswrapper[4809]: E1205 12:41:23.540458 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3103bf31-c6f8-494f-a75e-deb3cea6d915" containerName="mariadb-database-create" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.540465 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3103bf31-c6f8-494f-a75e-deb3cea6d915" containerName="mariadb-database-create" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.540628 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b95a840-4da9-4162-bf78-a89b0c0ce3ba" containerName="mariadb-account-create-update" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.540668 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3103bf31-c6f8-494f-a75e-deb3cea6d915" containerName="mariadb-database-create" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.542813 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.553381 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f894f5fcc-p99x8"] Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.577607 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-7r8q4"] Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.578587 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.586292 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-67s2j" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.586496 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.586604 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.605200 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7r8q4"] Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.629418 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-dns-svc\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.629488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-nb\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.629708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-sb\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.630500 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-config\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.630537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llcj7\" (UniqueName: \"kubernetes.io/projected/c094fae2-a65d-4d07-9621-76b469e2776f-kube-api-access-llcj7\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.732389 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpzkq\" (UniqueName: \"kubernetes.io/projected/325aafe3-ade7-4fbb-b031-53e17b3e6422-kube-api-access-mpzkq\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.732458 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-config\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.732482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llcj7\" (UniqueName: \"kubernetes.io/projected/c094fae2-a65d-4d07-9621-76b469e2776f-kube-api-access-llcj7\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.732593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-dns-svc\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.732624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-combined-ca-bundle\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.732688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-nb\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.732708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-scripts\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.733025 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-sb\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.733218 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-config-data\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.733511 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325aafe3-ade7-4fbb-b031-53e17b3e6422-logs\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.734029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-config\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.734045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-sb\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.734085 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-nb\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.734612 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-dns-svc\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.759019 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llcj7\" (UniqueName: \"kubernetes.io/projected/c094fae2-a65d-4d07-9621-76b469e2776f-kube-api-access-llcj7\") pod \"dnsmasq-dns-5f894f5fcc-p99x8\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.835840 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-combined-ca-bundle\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.835895 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-scripts\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.835986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-config-data\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.836088 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325aafe3-ade7-4fbb-b031-53e17b3e6422-logs\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.836112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpzkq\" (UniqueName: \"kubernetes.io/projected/325aafe3-ade7-4fbb-b031-53e17b3e6422-kube-api-access-mpzkq\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.836850 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325aafe3-ade7-4fbb-b031-53e17b3e6422-logs\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.840048 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-combined-ca-bundle\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.840492 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-config-data\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.855009 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpzkq\" (UniqueName: \"kubernetes.io/projected/325aafe3-ade7-4fbb-b031-53e17b3e6422-kube-api-access-mpzkq\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.856978 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-scripts\") pod \"placement-db-sync-7r8q4\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.866227 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:23 crc kubenswrapper[4809]: I1205 12:41:23.913379 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:24 crc kubenswrapper[4809]: I1205 12:41:24.349341 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f894f5fcc-p99x8"] Dec 05 12:41:24 crc kubenswrapper[4809]: I1205 12:41:24.429757 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7r8q4"] Dec 05 12:41:24 crc kubenswrapper[4809]: W1205 12:41:24.434182 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod325aafe3_ade7_4fbb_b031_53e17b3e6422.slice/crio-05ca0935b2c7cce72f8d07d0cb7b68eacad5246d08b9148628aec80d33a9675a WatchSource:0}: Error finding container 05ca0935b2c7cce72f8d07d0cb7b68eacad5246d08b9148628aec80d33a9675a: Status 404 returned error can't find the container with id 05ca0935b2c7cce72f8d07d0cb7b68eacad5246d08b9148628aec80d33a9675a Dec 05 12:41:25 crc kubenswrapper[4809]: I1205 12:41:25.226361 4809 generic.go:334] "Generic (PLEG): container finished" podID="c094fae2-a65d-4d07-9621-76b469e2776f" containerID="b179cd74f936ab579f93f15048d2ed355f9714ed2528e179285f8c5fd64ec384" exitCode=0 Dec 05 12:41:25 crc kubenswrapper[4809]: I1205 12:41:25.226417 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" event={"ID":"c094fae2-a65d-4d07-9621-76b469e2776f","Type":"ContainerDied","Data":"b179cd74f936ab579f93f15048d2ed355f9714ed2528e179285f8c5fd64ec384"} Dec 05 12:41:25 crc kubenswrapper[4809]: I1205 12:41:25.226866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" event={"ID":"c094fae2-a65d-4d07-9621-76b469e2776f","Type":"ContainerStarted","Data":"4837a9523f9293da731ee3194e8e3fe845f96bf68ad69a9c52ee762cbcdcc01e"} Dec 05 12:41:25 crc kubenswrapper[4809]: I1205 12:41:25.229738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7r8q4" event={"ID":"325aafe3-ade7-4fbb-b031-53e17b3e6422","Type":"ContainerStarted","Data":"6a95f60470fe38859395be6af8241e50657cff321bf12ca416c9af8b109948f5"} Dec 05 12:41:25 crc kubenswrapper[4809]: I1205 12:41:25.229772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7r8q4" event={"ID":"325aafe3-ade7-4fbb-b031-53e17b3e6422","Type":"ContainerStarted","Data":"05ca0935b2c7cce72f8d07d0cb7b68eacad5246d08b9148628aec80d33a9675a"} Dec 05 12:41:25 crc kubenswrapper[4809]: I1205 12:41:25.276385 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-7r8q4" podStartSLOduration=2.2763618709999998 podStartE2EDuration="2.276361871s" podCreationTimestamp="2025-12-05 12:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:41:25.261480253 +0000 UTC m=+5580.652456821" watchObservedRunningTime="2025-12-05 12:41:25.276361871 +0000 UTC m=+5580.667338429" Dec 05 12:41:26 crc kubenswrapper[4809]: I1205 12:41:26.241212 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" event={"ID":"c094fae2-a65d-4d07-9621-76b469e2776f","Type":"ContainerStarted","Data":"689fbc286b1035d92ae104840f2203382ea47c3a6da66e170e529aa06b318a53"} Dec 05 12:41:26 crc kubenswrapper[4809]: I1205 12:41:26.241544 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:26 crc kubenswrapper[4809]: I1205 12:41:26.243337 4809 generic.go:334] "Generic (PLEG): container finished" podID="325aafe3-ade7-4fbb-b031-53e17b3e6422" containerID="6a95f60470fe38859395be6af8241e50657cff321bf12ca416c9af8b109948f5" exitCode=0 Dec 05 12:41:26 crc kubenswrapper[4809]: I1205 12:41:26.243378 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7r8q4" event={"ID":"325aafe3-ade7-4fbb-b031-53e17b3e6422","Type":"ContainerDied","Data":"6a95f60470fe38859395be6af8241e50657cff321bf12ca416c9af8b109948f5"} Dec 05 12:41:26 crc kubenswrapper[4809]: I1205 12:41:26.265192 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" podStartSLOduration=3.26517475 podStartE2EDuration="3.26517475s" podCreationTimestamp="2025-12-05 12:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:41:26.264080021 +0000 UTC m=+5581.655056589" watchObservedRunningTime="2025-12-05 12:41:26.26517475 +0000 UTC m=+5581.656151308" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.574401 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.706312 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpzkq\" (UniqueName: \"kubernetes.io/projected/325aafe3-ade7-4fbb-b031-53e17b3e6422-kube-api-access-mpzkq\") pod \"325aafe3-ade7-4fbb-b031-53e17b3e6422\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.706406 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325aafe3-ade7-4fbb-b031-53e17b3e6422-logs\") pod \"325aafe3-ade7-4fbb-b031-53e17b3e6422\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.706482 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-config-data\") pod \"325aafe3-ade7-4fbb-b031-53e17b3e6422\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.706516 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-scripts\") pod \"325aafe3-ade7-4fbb-b031-53e17b3e6422\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.706832 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-combined-ca-bundle\") pod \"325aafe3-ade7-4fbb-b031-53e17b3e6422\" (UID: \"325aafe3-ade7-4fbb-b031-53e17b3e6422\") " Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.708186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/325aafe3-ade7-4fbb-b031-53e17b3e6422-logs" (OuterVolumeSpecName: "logs") pod "325aafe3-ade7-4fbb-b031-53e17b3e6422" (UID: "325aafe3-ade7-4fbb-b031-53e17b3e6422"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.718848 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-scripts" (OuterVolumeSpecName: "scripts") pod "325aafe3-ade7-4fbb-b031-53e17b3e6422" (UID: "325aafe3-ade7-4fbb-b031-53e17b3e6422"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.719079 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325aafe3-ade7-4fbb-b031-53e17b3e6422-kube-api-access-mpzkq" (OuterVolumeSpecName: "kube-api-access-mpzkq") pod "325aafe3-ade7-4fbb-b031-53e17b3e6422" (UID: "325aafe3-ade7-4fbb-b031-53e17b3e6422"). InnerVolumeSpecName "kube-api-access-mpzkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.732717 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "325aafe3-ade7-4fbb-b031-53e17b3e6422" (UID: "325aafe3-ade7-4fbb-b031-53e17b3e6422"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.733506 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-config-data" (OuterVolumeSpecName: "config-data") pod "325aafe3-ade7-4fbb-b031-53e17b3e6422" (UID: "325aafe3-ade7-4fbb-b031-53e17b3e6422"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.808480 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.808523 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpzkq\" (UniqueName: \"kubernetes.io/projected/325aafe3-ade7-4fbb-b031-53e17b3e6422-kube-api-access-mpzkq\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.808542 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325aafe3-ade7-4fbb-b031-53e17b3e6422-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.808554 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.808565 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325aafe3-ade7-4fbb-b031-53e17b3e6422-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.989696 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-84d8c96676-kw4cb"] Dec 05 12:41:27 crc kubenswrapper[4809]: E1205 12:41:27.990463 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325aafe3-ade7-4fbb-b031-53e17b3e6422" containerName="placement-db-sync" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.990484 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="325aafe3-ade7-4fbb-b031-53e17b3e6422" containerName="placement-db-sync" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.990753 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="325aafe3-ade7-4fbb-b031-53e17b3e6422" containerName="placement-db-sync" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.991959 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:27 crc kubenswrapper[4809]: I1205 12:41:27.999482 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84d8c96676-kw4cb"] Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.112783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws25z\" (UniqueName: \"kubernetes.io/projected/ffb727f4-12bb-40ea-b8ba-3bb33141727e-kube-api-access-ws25z\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.112836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb727f4-12bb-40ea-b8ba-3bb33141727e-logs\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.113038 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-combined-ca-bundle\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.113127 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-scripts\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.113157 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-config-data\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.214828 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws25z\" (UniqueName: \"kubernetes.io/projected/ffb727f4-12bb-40ea-b8ba-3bb33141727e-kube-api-access-ws25z\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.214882 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb727f4-12bb-40ea-b8ba-3bb33141727e-logs\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.214963 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-combined-ca-bundle\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.215001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-scripts\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.215025 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-config-data\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.216539 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffb727f4-12bb-40ea-b8ba-3bb33141727e-logs\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.219799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-config-data\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.220143 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-scripts\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.220593 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb727f4-12bb-40ea-b8ba-3bb33141727e-combined-ca-bundle\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.235120 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws25z\" (UniqueName: \"kubernetes.io/projected/ffb727f4-12bb-40ea-b8ba-3bb33141727e-kube-api-access-ws25z\") pod \"placement-84d8c96676-kw4cb\" (UID: \"ffb727f4-12bb-40ea-b8ba-3bb33141727e\") " pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.268427 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7r8q4" event={"ID":"325aafe3-ade7-4fbb-b031-53e17b3e6422","Type":"ContainerDied","Data":"05ca0935b2c7cce72f8d07d0cb7b68eacad5246d08b9148628aec80d33a9675a"} Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.268471 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05ca0935b2c7cce72f8d07d0cb7b68eacad5246d08b9148628aec80d33a9675a" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.268486 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7r8q4" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.311704 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:28 crc kubenswrapper[4809]: I1205 12:41:28.742222 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84d8c96676-kw4cb"] Dec 05 12:41:29 crc kubenswrapper[4809]: I1205 12:41:29.279768 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84d8c96676-kw4cb" event={"ID":"ffb727f4-12bb-40ea-b8ba-3bb33141727e","Type":"ContainerStarted","Data":"3fc94bbd9a64405e753a7733362b43be1f6c0b3f88ad42671b4e3fcb0c3435e6"} Dec 05 12:41:29 crc kubenswrapper[4809]: I1205 12:41:29.280060 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84d8c96676-kw4cb" event={"ID":"ffb727f4-12bb-40ea-b8ba-3bb33141727e","Type":"ContainerStarted","Data":"b4d8599efe5510ea91cec62dcd631cda5acbbf8d4572adf425b1483029180f79"} Dec 05 12:41:30 crc kubenswrapper[4809]: I1205 12:41:30.290128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84d8c96676-kw4cb" event={"ID":"ffb727f4-12bb-40ea-b8ba-3bb33141727e","Type":"ContainerStarted","Data":"4ac8e42d21bbed59302e04fb68f0f138b9401c25c22ef7e8f1f264240ae0186d"} Dec 05 12:41:30 crc kubenswrapper[4809]: I1205 12:41:30.291443 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:30 crc kubenswrapper[4809]: I1205 12:41:30.291466 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:41:30 crc kubenswrapper[4809]: I1205 12:41:30.317823 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-84d8c96676-kw4cb" podStartSLOduration=3.3177957510000002 podStartE2EDuration="3.317795751s" podCreationTimestamp="2025-12-05 12:41:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:41:30.311849242 +0000 UTC m=+5585.702825810" watchObservedRunningTime="2025-12-05 12:41:30.317795751 +0000 UTC m=+5585.708772309" Dec 05 12:41:33 crc kubenswrapper[4809]: I1205 12:41:33.868822 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:41:33 crc kubenswrapper[4809]: I1205 12:41:33.926128 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6787b8b8df-5q5ms"] Dec 05 12:41:33 crc kubenswrapper[4809]: I1205 12:41:33.926353 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" podUID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerName="dnsmasq-dns" containerID="cri-o://61469d6fc98bdbbf5f3b16a26a2cb939a1844e9ec99bfb61bb3847bc9db963f3" gracePeriod=10 Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.331198 4809 generic.go:334] "Generic (PLEG): container finished" podID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerID="61469d6fc98bdbbf5f3b16a26a2cb939a1844e9ec99bfb61bb3847bc9db963f3" exitCode=0 Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.331496 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" event={"ID":"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6","Type":"ContainerDied","Data":"61469d6fc98bdbbf5f3b16a26a2cb939a1844e9ec99bfb61bb3847bc9db963f3"} Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.414241 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.521740 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-sb\") pod \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.521858 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-config\") pod \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.521921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8pmk\" (UniqueName: \"kubernetes.io/projected/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-kube-api-access-g8pmk\") pod \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.521971 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-nb\") pod \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.522670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-dns-svc\") pod \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\" (UID: \"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6\") " Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.544755 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-kube-api-access-g8pmk" (OuterVolumeSpecName: "kube-api-access-g8pmk") pod "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" (UID: "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6"). InnerVolumeSpecName "kube-api-access-g8pmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.568447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" (UID: "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.570103 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" (UID: "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.571787 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-config" (OuterVolumeSpecName: "config") pod "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" (UID: "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.572945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" (UID: "8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.625029 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8pmk\" (UniqueName: \"kubernetes.io/projected/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-kube-api-access-g8pmk\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.625068 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.625091 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.625104 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.625116 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:41:34 crc kubenswrapper[4809]: I1205 12:41:34.936716 4809 scope.go:117] "RemoveContainer" containerID="5d0c3d20dc138d2c97c2246e5bf711d1f60dd7e71e0b8b8d4b4f63ef07534669" Dec 05 12:41:35 crc kubenswrapper[4809]: I1205 12:41:35.343510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" event={"ID":"8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6","Type":"ContainerDied","Data":"72db3dbe88e35211b49e63f580f2ee04e8740b363173aca51268f882155b703a"} Dec 05 12:41:35 crc kubenswrapper[4809]: I1205 12:41:35.343571 4809 scope.go:117] "RemoveContainer" containerID="61469d6fc98bdbbf5f3b16a26a2cb939a1844e9ec99bfb61bb3847bc9db963f3" Dec 05 12:41:35 crc kubenswrapper[4809]: I1205 12:41:35.343570 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6787b8b8df-5q5ms" Dec 05 12:41:35 crc kubenswrapper[4809]: I1205 12:41:35.365183 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6787b8b8df-5q5ms"] Dec 05 12:41:35 crc kubenswrapper[4809]: I1205 12:41:35.367041 4809 scope.go:117] "RemoveContainer" containerID="f432078dbf40f8bd74f484a7d063b1566e923ba530c0c3cd719112d6ba379fd1" Dec 05 12:41:35 crc kubenswrapper[4809]: I1205 12:41:35.372686 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6787b8b8df-5q5ms"] Dec 05 12:41:36 crc kubenswrapper[4809]: I1205 12:41:36.882297 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" path="/var/lib/kubelet/pods/8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6/volumes" Dec 05 12:41:59 crc kubenswrapper[4809]: I1205 12:41:59.429502 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:42:00 crc kubenswrapper[4809]: I1205 12:42:00.440085 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84d8c96676-kw4cb" Dec 05 12:42:14 crc kubenswrapper[4809]: I1205 12:42:14.046376 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:42:14 crc kubenswrapper[4809]: I1205 12:42:14.046942 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.399895 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lnxhz"] Dec 05 12:42:20 crc kubenswrapper[4809]: E1205 12:42:20.400862 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerName="init" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.400878 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerName="init" Dec 05 12:42:20 crc kubenswrapper[4809]: E1205 12:42:20.400896 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerName="dnsmasq-dns" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.400905 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerName="dnsmasq-dns" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.401164 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f7a4cca-bd30-4a4b-a841-d37c3e8cc4f6" containerName="dnsmasq-dns" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.401888 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.408357 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lnxhz"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.451282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d624929-bbe0-4674-8aa5-6415feaacb60-operator-scripts\") pod \"nova-api-db-create-lnxhz\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.451342 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvstn\" (UniqueName: \"kubernetes.io/projected/4d624929-bbe0-4674-8aa5-6415feaacb60-kube-api-access-lvstn\") pod \"nova-api-db-create-lnxhz\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.507397 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wmrmc"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.508590 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.553623 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ffkl\" (UniqueName: \"kubernetes.io/projected/dccc2924-a397-4c16-961b-26a510522a1c-kube-api-access-6ffkl\") pod \"nova-cell0-db-create-wmrmc\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.553754 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dccc2924-a397-4c16-961b-26a510522a1c-operator-scripts\") pod \"nova-cell0-db-create-wmrmc\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.553850 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d624929-bbe0-4674-8aa5-6415feaacb60-operator-scripts\") pod \"nova-api-db-create-lnxhz\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.553934 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvstn\" (UniqueName: \"kubernetes.io/projected/4d624929-bbe0-4674-8aa5-6415feaacb60-kube-api-access-lvstn\") pod \"nova-api-db-create-lnxhz\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.555115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d624929-bbe0-4674-8aa5-6415feaacb60-operator-scripts\") pod \"nova-api-db-create-lnxhz\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.556195 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wmrmc"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.577499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvstn\" (UniqueName: \"kubernetes.io/projected/4d624929-bbe0-4674-8aa5-6415feaacb60-kube-api-access-lvstn\") pod \"nova-api-db-create-lnxhz\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.615925 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-abbc-account-create-update-dql8l"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.617041 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.619780 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.628474 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-abbc-account-create-update-dql8l"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.656907 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ffkl\" (UniqueName: \"kubernetes.io/projected/dccc2924-a397-4c16-961b-26a510522a1c-kube-api-access-6ffkl\") pod \"nova-cell0-db-create-wmrmc\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.656956 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dccc2924-a397-4c16-961b-26a510522a1c-operator-scripts\") pod \"nova-cell0-db-create-wmrmc\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.657012 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57frj\" (UniqueName: \"kubernetes.io/projected/103c4a83-b894-416e-ab74-56ee1f206e44-kube-api-access-57frj\") pod \"nova-api-abbc-account-create-update-dql8l\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.657061 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103c4a83-b894-416e-ab74-56ee1f206e44-operator-scripts\") pod \"nova-api-abbc-account-create-update-dql8l\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.657822 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dccc2924-a397-4c16-961b-26a510522a1c-operator-scripts\") pod \"nova-cell0-db-create-wmrmc\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.678866 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ffkl\" (UniqueName: \"kubernetes.io/projected/dccc2924-a397-4c16-961b-26a510522a1c-kube-api-access-6ffkl\") pod \"nova-cell0-db-create-wmrmc\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.734977 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bdsxj"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.736120 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.759853 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bdsxj"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.762725 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57frj\" (UniqueName: \"kubernetes.io/projected/103c4a83-b894-416e-ab74-56ee1f206e44-kube-api-access-57frj\") pod \"nova-api-abbc-account-create-update-dql8l\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.762818 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103c4a83-b894-416e-ab74-56ee1f206e44-operator-scripts\") pod \"nova-api-abbc-account-create-update-dql8l\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.763597 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103c4a83-b894-416e-ab74-56ee1f206e44-operator-scripts\") pod \"nova-api-abbc-account-create-update-dql8l\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.778189 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.789237 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57frj\" (UniqueName: \"kubernetes.io/projected/103c4a83-b894-416e-ab74-56ee1f206e44-kube-api-access-57frj\") pod \"nova-api-abbc-account-create-update-dql8l\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.823931 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9e4c-account-create-update-wl2w2"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.827113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.830406 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.831145 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.840074 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9e4c-account-create-update-wl2w2"] Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.864608 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/443f06d4-d0db-4c99-8497-ffa1ffea10a8-operator-scripts\") pod \"nova-cell1-db-create-bdsxj\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.864943 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwbds\" (UniqueName: \"kubernetes.io/projected/443f06d4-d0db-4c99-8497-ffa1ffea10a8-kube-api-access-zwbds\") pod \"nova-cell1-db-create-bdsxj\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.949549 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.971099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwbds\" (UniqueName: \"kubernetes.io/projected/443f06d4-d0db-4c99-8497-ffa1ffea10a8-kube-api-access-zwbds\") pod \"nova-cell1-db-create-bdsxj\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.971164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-258tt\" (UniqueName: \"kubernetes.io/projected/e5074f64-b9a7-4675-b323-1a3af9beeb7f-kube-api-access-258tt\") pod \"nova-cell0-9e4c-account-create-update-wl2w2\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.971247 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5074f64-b9a7-4675-b323-1a3af9beeb7f-operator-scripts\") pod \"nova-cell0-9e4c-account-create-update-wl2w2\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.971270 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/443f06d4-d0db-4c99-8497-ffa1ffea10a8-operator-scripts\") pod \"nova-cell1-db-create-bdsxj\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:20 crc kubenswrapper[4809]: I1205 12:42:20.971967 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/443f06d4-d0db-4c99-8497-ffa1ffea10a8-operator-scripts\") pod \"nova-cell1-db-create-bdsxj\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.002497 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwbds\" (UniqueName: \"kubernetes.io/projected/443f06d4-d0db-4c99-8497-ffa1ffea10a8-kube-api-access-zwbds\") pod \"nova-cell1-db-create-bdsxj\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.026213 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-03e6-account-create-update-8vfsh"] Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.027423 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.029469 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.037265 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-03e6-account-create-update-8vfsh"] Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.067416 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.074322 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-258tt\" (UniqueName: \"kubernetes.io/projected/e5074f64-b9a7-4675-b323-1a3af9beeb7f-kube-api-access-258tt\") pod \"nova-cell0-9e4c-account-create-update-wl2w2\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.075134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5074f64-b9a7-4675-b323-1a3af9beeb7f-operator-scripts\") pod \"nova-cell0-9e4c-account-create-update-wl2w2\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.076400 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5074f64-b9a7-4675-b323-1a3af9beeb7f-operator-scripts\") pod \"nova-cell0-9e4c-account-create-update-wl2w2\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.098964 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-258tt\" (UniqueName: \"kubernetes.io/projected/e5074f64-b9a7-4675-b323-1a3af9beeb7f-kube-api-access-258tt\") pod \"nova-cell0-9e4c-account-create-update-wl2w2\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.182484 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbc6x\" (UniqueName: \"kubernetes.io/projected/e2d18b96-a42f-47fe-abcc-6a6ec909114e-kube-api-access-hbc6x\") pod \"nova-cell1-03e6-account-create-update-8vfsh\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.183026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d18b96-a42f-47fe-abcc-6a6ec909114e-operator-scripts\") pod \"nova-cell1-03e6-account-create-update-8vfsh\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.217378 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.285268 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbc6x\" (UniqueName: \"kubernetes.io/projected/e2d18b96-a42f-47fe-abcc-6a6ec909114e-kube-api-access-hbc6x\") pod \"nova-cell1-03e6-account-create-update-8vfsh\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.285361 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d18b96-a42f-47fe-abcc-6a6ec909114e-operator-scripts\") pod \"nova-cell1-03e6-account-create-update-8vfsh\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.286340 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d18b96-a42f-47fe-abcc-6a6ec909114e-operator-scripts\") pod \"nova-cell1-03e6-account-create-update-8vfsh\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.306771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbc6x\" (UniqueName: \"kubernetes.io/projected/e2d18b96-a42f-47fe-abcc-6a6ec909114e-kube-api-access-hbc6x\") pod \"nova-cell1-03e6-account-create-update-8vfsh\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.325483 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lnxhz"] Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.352080 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.410162 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wmrmc"] Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.592296 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bdsxj"] Dec 05 12:42:21 crc kubenswrapper[4809]: W1205 12:42:21.648226 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod443f06d4_d0db_4c99_8497_ffa1ffea10a8.slice/crio-12848c2fc864bd3542d3ec101ed0b3c951b898fc76512448346be8aefbf106fa WatchSource:0}: Error finding container 12848c2fc864bd3542d3ec101ed0b3c951b898fc76512448346be8aefbf106fa: Status 404 returned error can't find the container with id 12848c2fc864bd3542d3ec101ed0b3c951b898fc76512448346be8aefbf106fa Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.656425 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-abbc-account-create-update-dql8l"] Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.781131 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmrmc" event={"ID":"dccc2924-a397-4c16-961b-26a510522a1c","Type":"ContainerStarted","Data":"4f45e53c3e7b3f69dd1517f4f5e0b54123f4c25fcd1cc38c842e31da97f6f3e2"} Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.781185 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmrmc" event={"ID":"dccc2924-a397-4c16-961b-26a510522a1c","Type":"ContainerStarted","Data":"fbf8e5abe3acece5b2962616ffb168584e5cda489fa94c10387709ee391a7a6e"} Dec 05 12:42:21 crc kubenswrapper[4809]: W1205 12:42:21.798603 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5074f64_b9a7_4675_b323_1a3af9beeb7f.slice/crio-2e4b2cd99f6803e76cda51a2307e8ddc979bcf201aad7650edd1629cff478780 WatchSource:0}: Error finding container 2e4b2cd99f6803e76cda51a2307e8ddc979bcf201aad7650edd1629cff478780: Status 404 returned error can't find the container with id 2e4b2cd99f6803e76cda51a2307e8ddc979bcf201aad7650edd1629cff478780 Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.799723 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9e4c-account-create-update-wl2w2"] Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.807174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-abbc-account-create-update-dql8l" event={"ID":"103c4a83-b894-416e-ab74-56ee1f206e44","Type":"ContainerStarted","Data":"f9b215d0ab2f7834ffe46ae80a3e2c98fd0623e83547de6e469297d6c9d662ad"} Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.830296 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-wmrmc" podStartSLOduration=1.83027424 podStartE2EDuration="1.83027424s" podCreationTimestamp="2025-12-05 12:42:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:21.816097309 +0000 UTC m=+5637.207073877" watchObservedRunningTime="2025-12-05 12:42:21.83027424 +0000 UTC m=+5637.221250798" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.845542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lnxhz" event={"ID":"4d624929-bbe0-4674-8aa5-6415feaacb60","Type":"ContainerStarted","Data":"bcd812d0b2fdabc8e3b5af5ad8152747a2b26ce6247f8e4bd6507555f973fdba"} Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.845599 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lnxhz" event={"ID":"4d624929-bbe0-4674-8aa5-6415feaacb60","Type":"ContainerStarted","Data":"c599acc49176baea8d45cf3b57bd38d4b6c8941dc5cb4c9da6a14fe636e9dbf8"} Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.850607 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bdsxj" event={"ID":"443f06d4-d0db-4c99-8497-ffa1ffea10a8","Type":"ContainerStarted","Data":"12848c2fc864bd3542d3ec101ed0b3c951b898fc76512448346be8aefbf106fa"} Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.887308 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-lnxhz" podStartSLOduration=1.8872854650000002 podStartE2EDuration="1.887285465s" podCreationTimestamp="2025-12-05 12:42:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:21.87446605 +0000 UTC m=+5637.265442628" watchObservedRunningTime="2025-12-05 12:42:21.887285465 +0000 UTC m=+5637.278262023" Dec 05 12:42:21 crc kubenswrapper[4809]: I1205 12:42:21.899756 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-03e6-account-create-update-8vfsh"] Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.861248 4809 generic.go:334] "Generic (PLEG): container finished" podID="4d624929-bbe0-4674-8aa5-6415feaacb60" containerID="bcd812d0b2fdabc8e3b5af5ad8152747a2b26ce6247f8e4bd6507555f973fdba" exitCode=0 Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.861326 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lnxhz" event={"ID":"4d624929-bbe0-4674-8aa5-6415feaacb60","Type":"ContainerDied","Data":"bcd812d0b2fdabc8e3b5af5ad8152747a2b26ce6247f8e4bd6507555f973fdba"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.863874 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" event={"ID":"e5074f64-b9a7-4675-b323-1a3af9beeb7f","Type":"ContainerStarted","Data":"589f7452101fa440c93234a048b4e6cccdd2dd1b818482625b043d71467c6011"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.863920 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" event={"ID":"e5074f64-b9a7-4675-b323-1a3af9beeb7f","Type":"ContainerStarted","Data":"2e4b2cd99f6803e76cda51a2307e8ddc979bcf201aad7650edd1629cff478780"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.865447 4809 generic.go:334] "Generic (PLEG): container finished" podID="443f06d4-d0db-4c99-8497-ffa1ffea10a8" containerID="0456a8fda8ba44a040478dfa2f1796ce0e518eeb81de9e3bc1a22c6066761a23" exitCode=0 Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.865496 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bdsxj" event={"ID":"443f06d4-d0db-4c99-8497-ffa1ffea10a8","Type":"ContainerDied","Data":"0456a8fda8ba44a040478dfa2f1796ce0e518eeb81de9e3bc1a22c6066761a23"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.867003 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" event={"ID":"e2d18b96-a42f-47fe-abcc-6a6ec909114e","Type":"ContainerStarted","Data":"b5b53c981c27e44c124008d1ef0cad30dfac017d80a079817d0f62c4f2e3247b"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.867036 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" event={"ID":"e2d18b96-a42f-47fe-abcc-6a6ec909114e","Type":"ContainerStarted","Data":"7ff1ce17bacd45c2a6dcf5235bfbf271634fedefddbdc9a1495ec2496af56990"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.871893 4809 generic.go:334] "Generic (PLEG): container finished" podID="dccc2924-a397-4c16-961b-26a510522a1c" containerID="4f45e53c3e7b3f69dd1517f4f5e0b54123f4c25fcd1cc38c842e31da97f6f3e2" exitCode=0 Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.890960 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmrmc" event={"ID":"dccc2924-a397-4c16-961b-26a510522a1c","Type":"ContainerDied","Data":"4f45e53c3e7b3f69dd1517f4f5e0b54123f4c25fcd1cc38c842e31da97f6f3e2"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.891005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-abbc-account-create-update-dql8l" event={"ID":"103c4a83-b894-416e-ab74-56ee1f206e44","Type":"ContainerStarted","Data":"d3806d8e20c39f4ad1e496494af6b9d4ebd21cf1eab899f903cd4ef83b63be7b"} Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.896658 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" podStartSLOduration=2.896625256 podStartE2EDuration="2.896625256s" podCreationTimestamp="2025-12-05 12:42:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:22.892104445 +0000 UTC m=+5638.283081003" watchObservedRunningTime="2025-12-05 12:42:22.896625256 +0000 UTC m=+5638.287601804" Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.943023 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" podStartSLOduration=2.943000255 podStartE2EDuration="2.943000255s" podCreationTimestamp="2025-12-05 12:42:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:22.935611056 +0000 UTC m=+5638.326587614" watchObservedRunningTime="2025-12-05 12:42:22.943000255 +0000 UTC m=+5638.333976813" Dec 05 12:42:22 crc kubenswrapper[4809]: I1205 12:42:22.956052 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-abbc-account-create-update-dql8l" podStartSLOduration=2.9560279659999997 podStartE2EDuration="2.956027966s" podCreationTimestamp="2025-12-05 12:42:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:22.951086973 +0000 UTC m=+5638.342063551" watchObservedRunningTime="2025-12-05 12:42:22.956027966 +0000 UTC m=+5638.347004524" Dec 05 12:42:23 crc kubenswrapper[4809]: I1205 12:42:23.886457 4809 generic.go:334] "Generic (PLEG): container finished" podID="e5074f64-b9a7-4675-b323-1a3af9beeb7f" containerID="589f7452101fa440c93234a048b4e6cccdd2dd1b818482625b043d71467c6011" exitCode=0 Dec 05 12:42:23 crc kubenswrapper[4809]: I1205 12:42:23.886519 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" event={"ID":"e5074f64-b9a7-4675-b323-1a3af9beeb7f","Type":"ContainerDied","Data":"589f7452101fa440c93234a048b4e6cccdd2dd1b818482625b043d71467c6011"} Dec 05 12:42:23 crc kubenswrapper[4809]: I1205 12:42:23.889560 4809 generic.go:334] "Generic (PLEG): container finished" podID="e2d18b96-a42f-47fe-abcc-6a6ec909114e" containerID="b5b53c981c27e44c124008d1ef0cad30dfac017d80a079817d0f62c4f2e3247b" exitCode=0 Dec 05 12:42:23 crc kubenswrapper[4809]: I1205 12:42:23.889882 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" event={"ID":"e2d18b96-a42f-47fe-abcc-6a6ec909114e","Type":"ContainerDied","Data":"b5b53c981c27e44c124008d1ef0cad30dfac017d80a079817d0f62c4f2e3247b"} Dec 05 12:42:23 crc kubenswrapper[4809]: I1205 12:42:23.906837 4809 generic.go:334] "Generic (PLEG): container finished" podID="103c4a83-b894-416e-ab74-56ee1f206e44" containerID="d3806d8e20c39f4ad1e496494af6b9d4ebd21cf1eab899f903cd4ef83b63be7b" exitCode=0 Dec 05 12:42:23 crc kubenswrapper[4809]: I1205 12:42:23.906977 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-abbc-account-create-update-dql8l" event={"ID":"103c4a83-b894-416e-ab74-56ee1f206e44","Type":"ContainerDied","Data":"d3806d8e20c39f4ad1e496494af6b9d4ebd21cf1eab899f903cd4ef83b63be7b"} Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.348932 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.358539 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.363776 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472077 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dccc2924-a397-4c16-961b-26a510522a1c-operator-scripts\") pod \"dccc2924-a397-4c16-961b-26a510522a1c\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472151 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvstn\" (UniqueName: \"kubernetes.io/projected/4d624929-bbe0-4674-8aa5-6415feaacb60-kube-api-access-lvstn\") pod \"4d624929-bbe0-4674-8aa5-6415feaacb60\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472221 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/443f06d4-d0db-4c99-8497-ffa1ffea10a8-operator-scripts\") pod \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472289 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ffkl\" (UniqueName: \"kubernetes.io/projected/dccc2924-a397-4c16-961b-26a510522a1c-kube-api-access-6ffkl\") pod \"dccc2924-a397-4c16-961b-26a510522a1c\" (UID: \"dccc2924-a397-4c16-961b-26a510522a1c\") " Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472341 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d624929-bbe0-4674-8aa5-6415feaacb60-operator-scripts\") pod \"4d624929-bbe0-4674-8aa5-6415feaacb60\" (UID: \"4d624929-bbe0-4674-8aa5-6415feaacb60\") " Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472377 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwbds\" (UniqueName: \"kubernetes.io/projected/443f06d4-d0db-4c99-8497-ffa1ffea10a8-kube-api-access-zwbds\") pod \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\" (UID: \"443f06d4-d0db-4c99-8497-ffa1ffea10a8\") " Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472718 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dccc2924-a397-4c16-961b-26a510522a1c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dccc2924-a397-4c16-961b-26a510522a1c" (UID: "dccc2924-a397-4c16-961b-26a510522a1c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472760 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/443f06d4-d0db-4c99-8497-ffa1ffea10a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "443f06d4-d0db-4c99-8497-ffa1ffea10a8" (UID: "443f06d4-d0db-4c99-8497-ffa1ffea10a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.472809 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d624929-bbe0-4674-8aa5-6415feaacb60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d624929-bbe0-4674-8aa5-6415feaacb60" (UID: "4d624929-bbe0-4674-8aa5-6415feaacb60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.473261 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dccc2924-a397-4c16-961b-26a510522a1c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.473279 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/443f06d4-d0db-4c99-8497-ffa1ffea10a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.473288 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d624929-bbe0-4674-8aa5-6415feaacb60-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.477302 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/443f06d4-d0db-4c99-8497-ffa1ffea10a8-kube-api-access-zwbds" (OuterVolumeSpecName: "kube-api-access-zwbds") pod "443f06d4-d0db-4c99-8497-ffa1ffea10a8" (UID: "443f06d4-d0db-4c99-8497-ffa1ffea10a8"). InnerVolumeSpecName "kube-api-access-zwbds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.477393 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dccc2924-a397-4c16-961b-26a510522a1c-kube-api-access-6ffkl" (OuterVolumeSpecName: "kube-api-access-6ffkl") pod "dccc2924-a397-4c16-961b-26a510522a1c" (UID: "dccc2924-a397-4c16-961b-26a510522a1c"). InnerVolumeSpecName "kube-api-access-6ffkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.477591 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d624929-bbe0-4674-8aa5-6415feaacb60-kube-api-access-lvstn" (OuterVolumeSpecName: "kube-api-access-lvstn") pod "4d624929-bbe0-4674-8aa5-6415feaacb60" (UID: "4d624929-bbe0-4674-8aa5-6415feaacb60"). InnerVolumeSpecName "kube-api-access-lvstn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.574213 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ffkl\" (UniqueName: \"kubernetes.io/projected/dccc2924-a397-4c16-961b-26a510522a1c-kube-api-access-6ffkl\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.574251 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwbds\" (UniqueName: \"kubernetes.io/projected/443f06d4-d0db-4c99-8497-ffa1ffea10a8-kube-api-access-zwbds\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.574262 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvstn\" (UniqueName: \"kubernetes.io/projected/4d624929-bbe0-4674-8aa5-6415feaacb60-kube-api-access-lvstn\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.917071 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bdsxj" event={"ID":"443f06d4-d0db-4c99-8497-ffa1ffea10a8","Type":"ContainerDied","Data":"12848c2fc864bd3542d3ec101ed0b3c951b898fc76512448346be8aefbf106fa"} Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.917120 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12848c2fc864bd3542d3ec101ed0b3c951b898fc76512448346be8aefbf106fa" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.917087 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bdsxj" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.918336 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmrmc" event={"ID":"dccc2924-a397-4c16-961b-26a510522a1c","Type":"ContainerDied","Data":"fbf8e5abe3acece5b2962616ffb168584e5cda489fa94c10387709ee391a7a6e"} Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.918361 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmrmc" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.918375 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbf8e5abe3acece5b2962616ffb168584e5cda489fa94c10387709ee391a7a6e" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.919948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lnxhz" event={"ID":"4d624929-bbe0-4674-8aa5-6415feaacb60","Type":"ContainerDied","Data":"c599acc49176baea8d45cf3b57bd38d4b6c8941dc5cb4c9da6a14fe636e9dbf8"} Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.919977 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c599acc49176baea8d45cf3b57bd38d4b6c8941dc5cb4c9da6a14fe636e9dbf8" Dec 05 12:42:24 crc kubenswrapper[4809]: I1205 12:42:24.920029 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lnxhz" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.310257 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.393582 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbc6x\" (UniqueName: \"kubernetes.io/projected/e2d18b96-a42f-47fe-abcc-6a6ec909114e-kube-api-access-hbc6x\") pod \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.393843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d18b96-a42f-47fe-abcc-6a6ec909114e-operator-scripts\") pod \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\" (UID: \"e2d18b96-a42f-47fe-abcc-6a6ec909114e\") " Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.395749 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2d18b96-a42f-47fe-abcc-6a6ec909114e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e2d18b96-a42f-47fe-abcc-6a6ec909114e" (UID: "e2d18b96-a42f-47fe-abcc-6a6ec909114e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.402783 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d18b96-a42f-47fe-abcc-6a6ec909114e-kube-api-access-hbc6x" (OuterVolumeSpecName: "kube-api-access-hbc6x") pod "e2d18b96-a42f-47fe-abcc-6a6ec909114e" (UID: "e2d18b96-a42f-47fe-abcc-6a6ec909114e"). InnerVolumeSpecName "kube-api-access-hbc6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.467532 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.476737 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.501544 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2d18b96-a42f-47fe-abcc-6a6ec909114e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.501583 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbc6x\" (UniqueName: \"kubernetes.io/projected/e2d18b96-a42f-47fe-abcc-6a6ec909114e-kube-api-access-hbc6x\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.602892 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103c4a83-b894-416e-ab74-56ee1f206e44-operator-scripts\") pod \"103c4a83-b894-416e-ab74-56ee1f206e44\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.602959 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57frj\" (UniqueName: \"kubernetes.io/projected/103c4a83-b894-416e-ab74-56ee1f206e44-kube-api-access-57frj\") pod \"103c4a83-b894-416e-ab74-56ee1f206e44\" (UID: \"103c4a83-b894-416e-ab74-56ee1f206e44\") " Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.602977 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-258tt\" (UniqueName: \"kubernetes.io/projected/e5074f64-b9a7-4675-b323-1a3af9beeb7f-kube-api-access-258tt\") pod \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.603096 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5074f64-b9a7-4675-b323-1a3af9beeb7f-operator-scripts\") pod \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\" (UID: \"e5074f64-b9a7-4675-b323-1a3af9beeb7f\") " Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.604003 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5074f64-b9a7-4675-b323-1a3af9beeb7f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5074f64-b9a7-4675-b323-1a3af9beeb7f" (UID: "e5074f64-b9a7-4675-b323-1a3af9beeb7f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.605131 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/103c4a83-b894-416e-ab74-56ee1f206e44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "103c4a83-b894-416e-ab74-56ee1f206e44" (UID: "103c4a83-b894-416e-ab74-56ee1f206e44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.606111 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/103c4a83-b894-416e-ab74-56ee1f206e44-kube-api-access-57frj" (OuterVolumeSpecName: "kube-api-access-57frj") pod "103c4a83-b894-416e-ab74-56ee1f206e44" (UID: "103c4a83-b894-416e-ab74-56ee1f206e44"). InnerVolumeSpecName "kube-api-access-57frj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.606524 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5074f64-b9a7-4675-b323-1a3af9beeb7f-kube-api-access-258tt" (OuterVolumeSpecName: "kube-api-access-258tt") pod "e5074f64-b9a7-4675-b323-1a3af9beeb7f" (UID: "e5074f64-b9a7-4675-b323-1a3af9beeb7f"). InnerVolumeSpecName "kube-api-access-258tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.706174 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/103c4a83-b894-416e-ab74-56ee1f206e44-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.706216 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57frj\" (UniqueName: \"kubernetes.io/projected/103c4a83-b894-416e-ab74-56ee1f206e44-kube-api-access-57frj\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.706228 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-258tt\" (UniqueName: \"kubernetes.io/projected/e5074f64-b9a7-4675-b323-1a3af9beeb7f-kube-api-access-258tt\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.706237 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5074f64-b9a7-4675-b323-1a3af9beeb7f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.932295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" event={"ID":"e2d18b96-a42f-47fe-abcc-6a6ec909114e","Type":"ContainerDied","Data":"7ff1ce17bacd45c2a6dcf5235bfbf271634fedefddbdc9a1495ec2496af56990"} Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.932344 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ff1ce17bacd45c2a6dcf5235bfbf271634fedefddbdc9a1495ec2496af56990" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.932411 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-03e6-account-create-update-8vfsh" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.937124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-abbc-account-create-update-dql8l" event={"ID":"103c4a83-b894-416e-ab74-56ee1f206e44","Type":"ContainerDied","Data":"f9b215d0ab2f7834ffe46ae80a3e2c98fd0623e83547de6e469297d6c9d662ad"} Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.937323 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9b215d0ab2f7834ffe46ae80a3e2c98fd0623e83547de6e469297d6c9d662ad" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.937191 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbc-account-create-update-dql8l" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.939310 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" event={"ID":"e5074f64-b9a7-4675-b323-1a3af9beeb7f","Type":"ContainerDied","Data":"2e4b2cd99f6803e76cda51a2307e8ddc979bcf201aad7650edd1629cff478780"} Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.939356 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e4b2cd99f6803e76cda51a2307e8ddc979bcf201aad7650edd1629cff478780" Dec 05 12:42:25 crc kubenswrapper[4809]: I1205 12:42:25.939405 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9e4c-account-create-update-wl2w2" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.011369 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dpbn8"] Dec 05 12:42:31 crc kubenswrapper[4809]: E1205 12:42:31.012153 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d18b96-a42f-47fe-abcc-6a6ec909114e" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012165 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d18b96-a42f-47fe-abcc-6a6ec909114e" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: E1205 12:42:31.012178 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="103c4a83-b894-416e-ab74-56ee1f206e44" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012183 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="103c4a83-b894-416e-ab74-56ee1f206e44" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: E1205 12:42:31.012195 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5074f64-b9a7-4675-b323-1a3af9beeb7f" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012203 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5074f64-b9a7-4675-b323-1a3af9beeb7f" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: E1205 12:42:31.012224 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccc2924-a397-4c16-961b-26a510522a1c" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012230 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccc2924-a397-4c16-961b-26a510522a1c" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: E1205 12:42:31.012248 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d624929-bbe0-4674-8aa5-6415feaacb60" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012254 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d624929-bbe0-4674-8aa5-6415feaacb60" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: E1205 12:42:31.012266 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="443f06d4-d0db-4c99-8497-ffa1ffea10a8" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012271 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="443f06d4-d0db-4c99-8497-ffa1ffea10a8" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012422 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dccc2924-a397-4c16-961b-26a510522a1c" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012431 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="443f06d4-d0db-4c99-8497-ffa1ffea10a8" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012440 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="103c4a83-b894-416e-ab74-56ee1f206e44" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012457 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d624929-bbe0-4674-8aa5-6415feaacb60" containerName="mariadb-database-create" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012467 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d18b96-a42f-47fe-abcc-6a6ec909114e" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.012477 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5074f64-b9a7-4675-b323-1a3af9beeb7f" containerName="mariadb-account-create-update" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.013026 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.014837 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.014840 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-slkts" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.061099 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.072024 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dpbn8"] Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.195444 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-config-data\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.195500 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.195519 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-scripts\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.195556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgscj\" (UniqueName: \"kubernetes.io/projected/30302f31-4bc7-4fce-b7c5-61ad600ade36-kube-api-access-tgscj\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.297770 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-config-data\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.298095 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.298220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-scripts\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.298578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgscj\" (UniqueName: \"kubernetes.io/projected/30302f31-4bc7-4fce-b7c5-61ad600ade36-kube-api-access-tgscj\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.304591 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-scripts\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.304898 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.305080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-config-data\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.317920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgscj\" (UniqueName: \"kubernetes.io/projected/30302f31-4bc7-4fce-b7c5-61ad600ade36-kube-api-access-tgscj\") pod \"nova-cell0-conductor-db-sync-dpbn8\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.372791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.808236 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dpbn8"] Dec 05 12:42:31 crc kubenswrapper[4809]: W1205 12:42:31.813207 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30302f31_4bc7_4fce_b7c5_61ad600ade36.slice/crio-7f8e4353fc2bba5462f203826cac9af81d05bf0ad5999f8cfa4b90b48d52fb6f WatchSource:0}: Error finding container 7f8e4353fc2bba5462f203826cac9af81d05bf0ad5999f8cfa4b90b48d52fb6f: Status 404 returned error can't find the container with id 7f8e4353fc2bba5462f203826cac9af81d05bf0ad5999f8cfa4b90b48d52fb6f Dec 05 12:42:31 crc kubenswrapper[4809]: I1205 12:42:31.989324 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" event={"ID":"30302f31-4bc7-4fce-b7c5-61ad600ade36","Type":"ContainerStarted","Data":"7f8e4353fc2bba5462f203826cac9af81d05bf0ad5999f8cfa4b90b48d52fb6f"} Dec 05 12:42:32 crc kubenswrapper[4809]: I1205 12:42:32.998017 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" event={"ID":"30302f31-4bc7-4fce-b7c5-61ad600ade36","Type":"ContainerStarted","Data":"b0f8d2ae810086c971d66875a1ccf6202fde508960c503652f49a7d62fb56e69"} Dec 05 12:42:38 crc kubenswrapper[4809]: I1205 12:42:38.053554 4809 generic.go:334] "Generic (PLEG): container finished" podID="30302f31-4bc7-4fce-b7c5-61ad600ade36" containerID="b0f8d2ae810086c971d66875a1ccf6202fde508960c503652f49a7d62fb56e69" exitCode=0 Dec 05 12:42:38 crc kubenswrapper[4809]: I1205 12:42:38.053896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" event={"ID":"30302f31-4bc7-4fce-b7c5-61ad600ade36","Type":"ContainerDied","Data":"b0f8d2ae810086c971d66875a1ccf6202fde508960c503652f49a7d62fb56e69"} Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.365006 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.524329 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgscj\" (UniqueName: \"kubernetes.io/projected/30302f31-4bc7-4fce-b7c5-61ad600ade36-kube-api-access-tgscj\") pod \"30302f31-4bc7-4fce-b7c5-61ad600ade36\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.524724 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-scripts\") pod \"30302f31-4bc7-4fce-b7c5-61ad600ade36\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.524969 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-combined-ca-bundle\") pod \"30302f31-4bc7-4fce-b7c5-61ad600ade36\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.525851 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-config-data\") pod \"30302f31-4bc7-4fce-b7c5-61ad600ade36\" (UID: \"30302f31-4bc7-4fce-b7c5-61ad600ade36\") " Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.531063 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30302f31-4bc7-4fce-b7c5-61ad600ade36-kube-api-access-tgscj" (OuterVolumeSpecName: "kube-api-access-tgscj") pod "30302f31-4bc7-4fce-b7c5-61ad600ade36" (UID: "30302f31-4bc7-4fce-b7c5-61ad600ade36"). InnerVolumeSpecName "kube-api-access-tgscj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.532162 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-scripts" (OuterVolumeSpecName: "scripts") pod "30302f31-4bc7-4fce-b7c5-61ad600ade36" (UID: "30302f31-4bc7-4fce-b7c5-61ad600ade36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.550574 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30302f31-4bc7-4fce-b7c5-61ad600ade36" (UID: "30302f31-4bc7-4fce-b7c5-61ad600ade36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.551662 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-config-data" (OuterVolumeSpecName: "config-data") pod "30302f31-4bc7-4fce-b7c5-61ad600ade36" (UID: "30302f31-4bc7-4fce-b7c5-61ad600ade36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.627859 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.627897 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.627908 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgscj\" (UniqueName: \"kubernetes.io/projected/30302f31-4bc7-4fce-b7c5-61ad600ade36-kube-api-access-tgscj\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:39 crc kubenswrapper[4809]: I1205 12:42:39.627917 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30302f31-4bc7-4fce-b7c5-61ad600ade36-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.070771 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" event={"ID":"30302f31-4bc7-4fce-b7c5-61ad600ade36","Type":"ContainerDied","Data":"7f8e4353fc2bba5462f203826cac9af81d05bf0ad5999f8cfa4b90b48d52fb6f"} Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.070811 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f8e4353fc2bba5462f203826cac9af81d05bf0ad5999f8cfa4b90b48d52fb6f" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.070825 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dpbn8" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.167371 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:42:40 crc kubenswrapper[4809]: E1205 12:42:40.167870 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30302f31-4bc7-4fce-b7c5-61ad600ade36" containerName="nova-cell0-conductor-db-sync" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.167898 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="30302f31-4bc7-4fce-b7c5-61ad600ade36" containerName="nova-cell0-conductor-db-sync" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.168062 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="30302f31-4bc7-4fce-b7c5-61ad600ade36" containerName="nova-cell0-conductor-db-sync" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.168657 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.171363 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-slkts" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.171732 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.183278 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.339114 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.339181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.339405 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqfnr\" (UniqueName: \"kubernetes.io/projected/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-kube-api-access-sqfnr\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.441145 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.441234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.441315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqfnr\" (UniqueName: \"kubernetes.io/projected/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-kube-api-access-sqfnr\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.447176 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.449396 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.460001 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqfnr\" (UniqueName: \"kubernetes.io/projected/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-kube-api-access-sqfnr\") pod \"nova-cell0-conductor-0\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.498736 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:40 crc kubenswrapper[4809]: I1205 12:42:40.978604 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:42:41 crc kubenswrapper[4809]: I1205 12:42:41.082105 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"74c9484e-89ae-4abf-9cdc-2953bd18bf8c","Type":"ContainerStarted","Data":"03efaa4cf4a06b3d49dd9e39fe1c07c13e31919c10370d3409a7360ab9e2c631"} Dec 05 12:42:42 crc kubenswrapper[4809]: I1205 12:42:42.091413 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"74c9484e-89ae-4abf-9cdc-2953bd18bf8c","Type":"ContainerStarted","Data":"1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23"} Dec 05 12:42:42 crc kubenswrapper[4809]: I1205 12:42:42.091565 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:42 crc kubenswrapper[4809]: I1205 12:42:42.112869 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.112847691 podStartE2EDuration="2.112847691s" podCreationTimestamp="2025-12-05 12:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:42.105931295 +0000 UTC m=+5657.496907853" watchObservedRunningTime="2025-12-05 12:42:42.112847691 +0000 UTC m=+5657.503824259" Dec 05 12:42:44 crc kubenswrapper[4809]: I1205 12:42:44.046668 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:42:44 crc kubenswrapper[4809]: I1205 12:42:44.047013 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:42:50 crc kubenswrapper[4809]: I1205 12:42:50.526280 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 12:42:50 crc kubenswrapper[4809]: I1205 12:42:50.916804 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-snglg"] Dec 05 12:42:50 crc kubenswrapper[4809]: I1205 12:42:50.918611 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:50 crc kubenswrapper[4809]: I1205 12:42:50.920821 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 05 12:42:50 crc kubenswrapper[4809]: I1205 12:42:50.921185 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 05 12:42:50 crc kubenswrapper[4809]: I1205 12:42:50.926759 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-snglg"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.017435 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.018833 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.023841 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.037887 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.039549 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.039623 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-scripts\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.039687 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k45jd\" (UniqueName: \"kubernetes.io/projected/47917734-1cb9-4e2b-b016-36d5a57aa79c-kube-api-access-k45jd\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.039740 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.053758 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.055132 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.059781 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.079336 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143676 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k5pl\" (UniqueName: \"kubernetes.io/projected/3433b83d-e281-4415-9547-c1d2868331fa-kube-api-access-2k5pl\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143725 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143763 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-scripts\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143806 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k45jd\" (UniqueName: \"kubernetes.io/projected/47917734-1cb9-4e2b-b016-36d5a57aa79c-kube-api-access-k45jd\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143826 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143862 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143896 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lxvk\" (UniqueName: \"kubernetes.io/projected/bc68a01f-217d-4d14-bf85-05e3250c3a8e-kube-api-access-8lxvk\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143925 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3433b83d-e281-4415-9547-c1d2868331fa-logs\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143955 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.143980 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-config-data\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.144010 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.153327 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-scripts\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.153530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.154914 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.154954 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.161597 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.166016 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.187660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k45jd\" (UniqueName: \"kubernetes.io/projected/47917734-1cb9-4e2b-b016-36d5a57aa79c-kube-api-access-k45jd\") pod \"nova-cell0-cell-mapping-snglg\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.200399 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245573 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3433b83d-e281-4415-9547-c1d2868331fa-logs\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245696 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-config-data\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245749 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k5pl\" (UniqueName: \"kubernetes.io/projected/3433b83d-e281-4415-9547-c1d2868331fa-kube-api-access-2k5pl\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245773 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245900 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.245916 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lxvk\" (UniqueName: \"kubernetes.io/projected/bc68a01f-217d-4d14-bf85-05e3250c3a8e-kube-api-access-8lxvk\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.252953 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3433b83d-e281-4415-9547-c1d2868331fa-logs\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.260202 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.262323 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.283655 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-config-data\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.284098 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.285446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k5pl\" (UniqueName: \"kubernetes.io/projected/3433b83d-e281-4415-9547-c1d2868331fa-kube-api-access-2k5pl\") pod \"nova-api-0\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.285494 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bd474b5b7-sm5l9"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.286960 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.294616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lxvk\" (UniqueName: \"kubernetes.io/projected/bc68a01f-217d-4d14-bf85-05e3250c3a8e-kube-api-access-8lxvk\") pod \"nova-cell1-novncproxy-0\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.340718 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.342777 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.344357 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.354278 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.358840 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h6x7\" (UniqueName: \"kubernetes.io/projected/c5182106-7776-4589-877b-2855de564de7-kube-api-access-6h6x7\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.359212 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.359366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-config-data\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.359671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5182106-7776-4589-877b-2855de564de7-logs\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.367978 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd474b5b7-sm5l9"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.381382 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.410308 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523490 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-config-data\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-dns-svc\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h6x7\" (UniqueName: \"kubernetes.io/projected/c5182106-7776-4589-877b-2855de564de7-kube-api-access-6h6x7\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523658 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpgfg\" (UniqueName: \"kubernetes.io/projected/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-kube-api-access-vpgfg\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-config\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523734 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523806 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-config-data\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523881 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.523913 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqglw\" (UniqueName: \"kubernetes.io/projected/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-kube-api-access-mqglw\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.524007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.524043 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5182106-7776-4589-877b-2855de564de7-logs\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.524086 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.539780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5182106-7776-4589-877b-2855de564de7-logs\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.546587 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.568174 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h6x7\" (UniqueName: \"kubernetes.io/projected/c5182106-7776-4589-877b-2855de564de7-kube-api-access-6h6x7\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.568291 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-config-data\") pod \"nova-metadata-0\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633335 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-config-data\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633667 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-dns-svc\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpgfg\" (UniqueName: \"kubernetes.io/projected/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-kube-api-access-vpgfg\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633741 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-config\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633815 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633843 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqglw\" (UniqueName: \"kubernetes.io/projected/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-kube-api-access-mqglw\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633866 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.633901 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.635564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.635564 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-dns-svc\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.635753 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-config\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.635780 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.637912 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-config-data\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.638036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.670862 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpgfg\" (UniqueName: \"kubernetes.io/projected/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-kube-api-access-vpgfg\") pod \"dnsmasq-dns-5bd474b5b7-sm5l9\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.671114 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqglw\" (UniqueName: \"kubernetes.io/projected/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-kube-api-access-mqglw\") pod \"nova-scheduler-0\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " pod="openstack/nova-scheduler-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.735453 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:42:51 crc kubenswrapper[4809]: I1205 12:42:51.825844 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:51.857675 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:51.951760 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-snglg"] Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.072854 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.133219 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b6jfm"] Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.136382 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.140994 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.141165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.147733 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:42:52 crc kubenswrapper[4809]: W1205 12:42:52.155714 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc68a01f_217d_4d14_bf85_05e3250c3a8e.slice/crio-5bd258579ab0cf63dbc2d055f2960d5b5bfb05ed004f40d634e8ca70599ea62d WatchSource:0}: Error finding container 5bd258579ab0cf63dbc2d055f2960d5b5bfb05ed004f40d634e8ca70599ea62d: Status 404 returned error can't find the container with id 5bd258579ab0cf63dbc2d055f2960d5b5bfb05ed004f40d634e8ca70599ea62d Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.162199 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b6jfm"] Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.233685 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3433b83d-e281-4415-9547-c1d2868331fa","Type":"ContainerStarted","Data":"f5817cbaec2ed5c23d51c5e7af2b278a8b3adba1acad87de227d02ae70a9f2d5"} Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.236071 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-snglg" event={"ID":"47917734-1cb9-4e2b-b016-36d5a57aa79c","Type":"ContainerStarted","Data":"88013cfb304725872ba8f39d1287ad224065d6e1ffe1ded2a37d03ae03693e1d"} Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.236154 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-snglg" event={"ID":"47917734-1cb9-4e2b-b016-36d5a57aa79c","Type":"ContainerStarted","Data":"f403cdf0f1cc71ffcb175edbb5bb9fabd2d4ffe816b32af3d0e7ed87cb3d3c48"} Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.240714 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bc68a01f-217d-4d14-bf85-05e3250c3a8e","Type":"ContainerStarted","Data":"5bd258579ab0cf63dbc2d055f2960d5b5bfb05ed004f40d634e8ca70599ea62d"} Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.259483 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnmjv\" (UniqueName: \"kubernetes.io/projected/fb2a36fb-5edb-4b55-bccf-df95e3d73500-kube-api-access-cnmjv\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.259647 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-scripts\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.259720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.259932 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-config-data\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.263434 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-snglg" podStartSLOduration=2.263419131 podStartE2EDuration="2.263419131s" podCreationTimestamp="2025-12-05 12:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:52.25889212 +0000 UTC m=+5667.649868678" watchObservedRunningTime="2025-12-05 12:42:52.263419131 +0000 UTC m=+5667.654395689" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.365316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnmjv\" (UniqueName: \"kubernetes.io/projected/fb2a36fb-5edb-4b55-bccf-df95e3d73500-kube-api-access-cnmjv\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.366089 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-scripts\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.366569 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.366648 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-config-data\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.370262 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.371545 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-scripts\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.374362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-config-data\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.384484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnmjv\" (UniqueName: \"kubernetes.io/projected/fb2a36fb-5edb-4b55-bccf-df95e3d73500-kube-api-access-cnmjv\") pod \"nova-cell1-conductor-db-sync-b6jfm\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.591448 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.926679 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd474b5b7-sm5l9"] Dec 05 12:42:52 crc kubenswrapper[4809]: W1205 12:42:52.932313 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56fb7d0e_51fd_42c6_a4e4_2d8953d3a4d1.slice/crio-d76b32bdd6c817a87f167dce0b662c529f2b0efe8f7bd6465e9481064172e860 WatchSource:0}: Error finding container d76b32bdd6c817a87f167dce0b662c529f2b0efe8f7bd6465e9481064172e860: Status 404 returned error can't find the container with id d76b32bdd6c817a87f167dce0b662c529f2b0efe8f7bd6465e9481064172e860 Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.945050 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:42:52 crc kubenswrapper[4809]: W1205 12:42:52.950740 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60eff64d_276b_41c5_ab5f_eda25d2a0c6c.slice/crio-a3869bae9a5581f98b7c5a00ca0440ab0179d19ae922757f2c43a3beb748bb24 WatchSource:0}: Error finding container a3869bae9a5581f98b7c5a00ca0440ab0179d19ae922757f2c43a3beb748bb24: Status 404 returned error can't find the container with id a3869bae9a5581f98b7c5a00ca0440ab0179d19ae922757f2c43a3beb748bb24 Dec 05 12:42:52 crc kubenswrapper[4809]: I1205 12:42:52.954730 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.136372 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b6jfm"] Dec 05 12:42:53 crc kubenswrapper[4809]: W1205 12:42:53.140854 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb2a36fb_5edb_4b55_bccf_df95e3d73500.slice/crio-9970189b34efbb76051c2de7efe20a0ec03ecc75825eb0f557aa5a512c8d9800 WatchSource:0}: Error finding container 9970189b34efbb76051c2de7efe20a0ec03ecc75825eb0f557aa5a512c8d9800: Status 404 returned error can't find the container with id 9970189b34efbb76051c2de7efe20a0ec03ecc75825eb0f557aa5a512c8d9800 Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.250783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bc68a01f-217d-4d14-bf85-05e3250c3a8e","Type":"ContainerStarted","Data":"d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.259063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3433b83d-e281-4415-9547-c1d2868331fa","Type":"ContainerStarted","Data":"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.259108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3433b83d-e281-4415-9547-c1d2868331fa","Type":"ContainerStarted","Data":"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.260888 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" event={"ID":"fb2a36fb-5edb-4b55-bccf-df95e3d73500","Type":"ContainerStarted","Data":"9970189b34efbb76051c2de7efe20a0ec03ecc75825eb0f557aa5a512c8d9800"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.262316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5182106-7776-4589-877b-2855de564de7","Type":"ContainerStarted","Data":"d1780c19b16939eeaaa8bc199ed0de29a951741ba1343c869efa054a59398bb7"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.262348 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5182106-7776-4589-877b-2855de564de7","Type":"ContainerStarted","Data":"2ecfd02123498532083d30f016e18d7f17052084c3b7f3dc73028b8a8fb9a063"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.265601 4809 generic.go:334] "Generic (PLEG): container finished" podID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerID="e96bafdbb4545cd249866567632a0a6220682b58d7e7c3aec925645c302b8ea3" exitCode=0 Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.265689 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" event={"ID":"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1","Type":"ContainerDied","Data":"e96bafdbb4545cd249866567632a0a6220682b58d7e7c3aec925645c302b8ea3"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.265710 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" event={"ID":"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1","Type":"ContainerStarted","Data":"d76b32bdd6c817a87f167dce0b662c529f2b0efe8f7bd6465e9481064172e860"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.274567 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60eff64d-276b-41c5-ab5f-eda25d2a0c6c","Type":"ContainerStarted","Data":"8cc99fa11e6dc7ae43979c4a85c0b43ba35752a89ca4360a07c64982202725d9"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.274648 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60eff64d-276b-41c5-ab5f-eda25d2a0c6c","Type":"ContainerStarted","Data":"a3869bae9a5581f98b7c5a00ca0440ab0179d19ae922757f2c43a3beb748bb24"} Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.277549 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.277535202 podStartE2EDuration="2.277535202s" podCreationTimestamp="2025-12-05 12:42:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:53.274781937 +0000 UTC m=+5668.665758495" watchObservedRunningTime="2025-12-05 12:42:53.277535202 +0000 UTC m=+5668.668511760" Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.319889 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.319868411 podStartE2EDuration="3.319868411s" podCreationTimestamp="2025-12-05 12:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:53.314522167 +0000 UTC m=+5668.705498715" watchObservedRunningTime="2025-12-05 12:42:53.319868411 +0000 UTC m=+5668.710844979" Dec 05 12:42:53 crc kubenswrapper[4809]: I1205 12:42:53.341974 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.341948606 podStartE2EDuration="2.341948606s" podCreationTimestamp="2025-12-05 12:42:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:53.333869688 +0000 UTC m=+5668.724846236" watchObservedRunningTime="2025-12-05 12:42:53.341948606 +0000 UTC m=+5668.732925164" Dec 05 12:42:54 crc kubenswrapper[4809]: I1205 12:42:54.288820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" event={"ID":"fb2a36fb-5edb-4b55-bccf-df95e3d73500","Type":"ContainerStarted","Data":"5f97f36da62aa79461a6d927b851869fe25656baf29470378972b7cc29f31157"} Dec 05 12:42:54 crc kubenswrapper[4809]: I1205 12:42:54.294081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5182106-7776-4589-877b-2855de564de7","Type":"ContainerStarted","Data":"e46ecbca15b989fbf4b4d0b601955eb37da770af3bdb56b624aa574edd1b6d4e"} Dec 05 12:42:54 crc kubenswrapper[4809]: I1205 12:42:54.296790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" event={"ID":"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1","Type":"ContainerStarted","Data":"7be44412e11c190acedf5a4d6d50042b56c1a3311699c368a608ccea05ed1abe"} Dec 05 12:42:54 crc kubenswrapper[4809]: I1205 12:42:54.334161 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" podStartSLOduration=2.334139824 podStartE2EDuration="2.334139824s" podCreationTimestamp="2025-12-05 12:42:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:54.306343426 +0000 UTC m=+5669.697319994" watchObservedRunningTime="2025-12-05 12:42:54.334139824 +0000 UTC m=+5669.725116382" Dec 05 12:42:54 crc kubenswrapper[4809]: I1205 12:42:54.338571 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.338548263 podStartE2EDuration="3.338548263s" podCreationTimestamp="2025-12-05 12:42:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:54.322747508 +0000 UTC m=+5669.713724066" watchObservedRunningTime="2025-12-05 12:42:54.338548263 +0000 UTC m=+5669.729524821" Dec 05 12:42:54 crc kubenswrapper[4809]: I1205 12:42:54.353779 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" podStartSLOduration=3.353753792 podStartE2EDuration="3.353753792s" podCreationTimestamp="2025-12-05 12:42:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:42:54.345664525 +0000 UTC m=+5669.736641093" watchObservedRunningTime="2025-12-05 12:42:54.353753792 +0000 UTC m=+5669.744730350" Dec 05 12:42:55 crc kubenswrapper[4809]: I1205 12:42:55.304214 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:42:56 crc kubenswrapper[4809]: I1205 12:42:56.382307 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:42:56 crc kubenswrapper[4809]: I1205 12:42:56.737039 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:42:56 crc kubenswrapper[4809]: I1205 12:42:56.738075 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:42:56 crc kubenswrapper[4809]: I1205 12:42:56.858744 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 12:42:57 crc kubenswrapper[4809]: I1205 12:42:57.321247 4809 generic.go:334] "Generic (PLEG): container finished" podID="47917734-1cb9-4e2b-b016-36d5a57aa79c" containerID="88013cfb304725872ba8f39d1287ad224065d6e1ffe1ded2a37d03ae03693e1d" exitCode=0 Dec 05 12:42:57 crc kubenswrapper[4809]: I1205 12:42:57.321330 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-snglg" event={"ID":"47917734-1cb9-4e2b-b016-36d5a57aa79c","Type":"ContainerDied","Data":"88013cfb304725872ba8f39d1287ad224065d6e1ffe1ded2a37d03ae03693e1d"} Dec 05 12:42:57 crc kubenswrapper[4809]: I1205 12:42:57.323391 4809 generic.go:334] "Generic (PLEG): container finished" podID="fb2a36fb-5edb-4b55-bccf-df95e3d73500" containerID="5f97f36da62aa79461a6d927b851869fe25656baf29470378972b7cc29f31157" exitCode=0 Dec 05 12:42:57 crc kubenswrapper[4809]: I1205 12:42:57.323538 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" event={"ID":"fb2a36fb-5edb-4b55-bccf-df95e3d73500","Type":"ContainerDied","Data":"5f97f36da62aa79461a6d927b851869fe25656baf29470378972b7cc29f31157"} Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.823438 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.829048 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.917659 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-scripts\") pod \"47917734-1cb9-4e2b-b016-36d5a57aa79c\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.917771 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data\") pod \"47917734-1cb9-4e2b-b016-36d5a57aa79c\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.917883 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k45jd\" (UniqueName: \"kubernetes.io/projected/47917734-1cb9-4e2b-b016-36d5a57aa79c-kube-api-access-k45jd\") pod \"47917734-1cb9-4e2b-b016-36d5a57aa79c\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.917903 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-combined-ca-bundle\") pod \"47917734-1cb9-4e2b-b016-36d5a57aa79c\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.923247 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-scripts" (OuterVolumeSpecName: "scripts") pod "47917734-1cb9-4e2b-b016-36d5a57aa79c" (UID: "47917734-1cb9-4e2b-b016-36d5a57aa79c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.923254 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47917734-1cb9-4e2b-b016-36d5a57aa79c-kube-api-access-k45jd" (OuterVolumeSpecName: "kube-api-access-k45jd") pod "47917734-1cb9-4e2b-b016-36d5a57aa79c" (UID: "47917734-1cb9-4e2b-b016-36d5a57aa79c"). InnerVolumeSpecName "kube-api-access-k45jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:58 crc kubenswrapper[4809]: E1205 12:42:58.944527 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data podName:47917734-1cb9-4e2b-b016-36d5a57aa79c nodeName:}" failed. No retries permitted until 2025-12-05 12:42:59.444499615 +0000 UTC m=+5674.835476173 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data") pod "47917734-1cb9-4e2b-b016-36d5a57aa79c" (UID: "47917734-1cb9-4e2b-b016-36d5a57aa79c") : error deleting /var/lib/kubelet/pods/47917734-1cb9-4e2b-b016-36d5a57aa79c/volume-subpaths: remove /var/lib/kubelet/pods/47917734-1cb9-4e2b-b016-36d5a57aa79c/volume-subpaths: no such file or directory Dec 05 12:42:58 crc kubenswrapper[4809]: I1205 12:42:58.947565 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47917734-1cb9-4e2b-b016-36d5a57aa79c" (UID: "47917734-1cb9-4e2b-b016-36d5a57aa79c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.019842 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnmjv\" (UniqueName: \"kubernetes.io/projected/fb2a36fb-5edb-4b55-bccf-df95e3d73500-kube-api-access-cnmjv\") pod \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.019938 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-scripts\") pod \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.019981 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-combined-ca-bundle\") pod \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.020106 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-config-data\") pod \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\" (UID: \"fb2a36fb-5edb-4b55-bccf-df95e3d73500\") " Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.020485 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k45jd\" (UniqueName: \"kubernetes.io/projected/47917734-1cb9-4e2b-b016-36d5a57aa79c-kube-api-access-k45jd\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.020501 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.020514 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.023418 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb2a36fb-5edb-4b55-bccf-df95e3d73500-kube-api-access-cnmjv" (OuterVolumeSpecName: "kube-api-access-cnmjv") pod "fb2a36fb-5edb-4b55-bccf-df95e3d73500" (UID: "fb2a36fb-5edb-4b55-bccf-df95e3d73500"). InnerVolumeSpecName "kube-api-access-cnmjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.023651 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-scripts" (OuterVolumeSpecName: "scripts") pod "fb2a36fb-5edb-4b55-bccf-df95e3d73500" (UID: "fb2a36fb-5edb-4b55-bccf-df95e3d73500"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.044272 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-config-data" (OuterVolumeSpecName: "config-data") pod "fb2a36fb-5edb-4b55-bccf-df95e3d73500" (UID: "fb2a36fb-5edb-4b55-bccf-df95e3d73500"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.046676 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb2a36fb-5edb-4b55-bccf-df95e3d73500" (UID: "fb2a36fb-5edb-4b55-bccf-df95e3d73500"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.122783 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.122836 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnmjv\" (UniqueName: \"kubernetes.io/projected/fb2a36fb-5edb-4b55-bccf-df95e3d73500-kube-api-access-cnmjv\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.122849 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.122861 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2a36fb-5edb-4b55-bccf-df95e3d73500-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.344897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-snglg" event={"ID":"47917734-1cb9-4e2b-b016-36d5a57aa79c","Type":"ContainerDied","Data":"f403cdf0f1cc71ffcb175edbb5bb9fabd2d4ffe816b32af3d0e7ed87cb3d3c48"} Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.344945 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f403cdf0f1cc71ffcb175edbb5bb9fabd2d4ffe816b32af3d0e7ed87cb3d3c48" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.344965 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-snglg" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.347530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" event={"ID":"fb2a36fb-5edb-4b55-bccf-df95e3d73500","Type":"ContainerDied","Data":"9970189b34efbb76051c2de7efe20a0ec03ecc75825eb0f557aa5a512c8d9800"} Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.347574 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9970189b34efbb76051c2de7efe20a0ec03ecc75825eb0f557aa5a512c8d9800" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.347572 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b6jfm" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.437509 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:42:59 crc kubenswrapper[4809]: E1205 12:42:59.438185 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2a36fb-5edb-4b55-bccf-df95e3d73500" containerName="nova-cell1-conductor-db-sync" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.438215 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2a36fb-5edb-4b55-bccf-df95e3d73500" containerName="nova-cell1-conductor-db-sync" Dec 05 12:42:59 crc kubenswrapper[4809]: E1205 12:42:59.438249 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47917734-1cb9-4e2b-b016-36d5a57aa79c" containerName="nova-manage" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.438257 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="47917734-1cb9-4e2b-b016-36d5a57aa79c" containerName="nova-manage" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.438431 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb2a36fb-5edb-4b55-bccf-df95e3d73500" containerName="nova-cell1-conductor-db-sync" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.438458 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="47917734-1cb9-4e2b-b016-36d5a57aa79c" containerName="nova-manage" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.439269 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.441587 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.451005 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.531035 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data\") pod \"47917734-1cb9-4e2b-b016-36d5a57aa79c\" (UID: \"47917734-1cb9-4e2b-b016-36d5a57aa79c\") " Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.535243 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data" (OuterVolumeSpecName: "config-data") pod "47917734-1cb9-4e2b-b016-36d5a57aa79c" (UID: "47917734-1cb9-4e2b-b016-36d5a57aa79c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.614346 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.614728 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-log" containerID="cri-o://1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389" gracePeriod=30 Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.614811 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-api" containerID="cri-o://ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388" gracePeriod=30 Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.627494 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.627879 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="60eff64d-276b-41c5-ab5f-eda25d2a0c6c" containerName="nova-scheduler-scheduler" containerID="cri-o://8cc99fa11e6dc7ae43979c4a85c0b43ba35752a89ca4360a07c64982202725d9" gracePeriod=30 Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.633818 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.633900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.633976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5d8p\" (UniqueName: \"kubernetes.io/projected/c097deb0-a548-4b13-a41f-68fa3b98074a-kube-api-access-j5d8p\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.634104 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47917734-1cb9-4e2b-b016-36d5a57aa79c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.712296 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.714292 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-log" containerID="cri-o://d1780c19b16939eeaaa8bc199ed0de29a951741ba1343c869efa054a59398bb7" gracePeriod=30 Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.714509 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-metadata" containerID="cri-o://e46ecbca15b989fbf4b4d0b601955eb37da770af3bdb56b624aa574edd1b6d4e" gracePeriod=30 Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.735250 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.735308 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.735394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5d8p\" (UniqueName: \"kubernetes.io/projected/c097deb0-a548-4b13-a41f-68fa3b98074a-kube-api-access-j5d8p\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.740737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.741874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.757346 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5d8p\" (UniqueName: \"kubernetes.io/projected/c097deb0-a548-4b13-a41f-68fa3b98074a-kube-api-access-j5d8p\") pod \"nova-cell1-conductor-0\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:42:59 crc kubenswrapper[4809]: I1205 12:42:59.763377 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.168507 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.346286 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-combined-ca-bundle\") pod \"3433b83d-e281-4415-9547-c1d2868331fa\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.346893 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k5pl\" (UniqueName: \"kubernetes.io/projected/3433b83d-e281-4415-9547-c1d2868331fa-kube-api-access-2k5pl\") pod \"3433b83d-e281-4415-9547-c1d2868331fa\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.346947 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3433b83d-e281-4415-9547-c1d2868331fa-logs\") pod \"3433b83d-e281-4415-9547-c1d2868331fa\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.347036 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-config-data\") pod \"3433b83d-e281-4415-9547-c1d2868331fa\" (UID: \"3433b83d-e281-4415-9547-c1d2868331fa\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.348025 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3433b83d-e281-4415-9547-c1d2868331fa-logs" (OuterVolumeSpecName: "logs") pod "3433b83d-e281-4415-9547-c1d2868331fa" (UID: "3433b83d-e281-4415-9547-c1d2868331fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.363432 4809 generic.go:334] "Generic (PLEG): container finished" podID="60eff64d-276b-41c5-ab5f-eda25d2a0c6c" containerID="8cc99fa11e6dc7ae43979c4a85c0b43ba35752a89ca4360a07c64982202725d9" exitCode=0 Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.363491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60eff64d-276b-41c5-ab5f-eda25d2a0c6c","Type":"ContainerDied","Data":"8cc99fa11e6dc7ae43979c4a85c0b43ba35752a89ca4360a07c64982202725d9"} Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.364732 4809 generic.go:334] "Generic (PLEG): container finished" podID="3433b83d-e281-4415-9547-c1d2868331fa" containerID="ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388" exitCode=0 Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.364753 4809 generic.go:334] "Generic (PLEG): container finished" podID="3433b83d-e281-4415-9547-c1d2868331fa" containerID="1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389" exitCode=143 Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.364790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3433b83d-e281-4415-9547-c1d2868331fa","Type":"ContainerDied","Data":"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388"} Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.364809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3433b83d-e281-4415-9547-c1d2868331fa","Type":"ContainerDied","Data":"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389"} Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.364819 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3433b83d-e281-4415-9547-c1d2868331fa","Type":"ContainerDied","Data":"f5817cbaec2ed5c23d51c5e7af2b278a8b3adba1acad87de227d02ae70a9f2d5"} Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.364833 4809 scope.go:117] "RemoveContainer" containerID="ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.364948 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.367809 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.369230 4809 generic.go:334] "Generic (PLEG): container finished" podID="c5182106-7776-4589-877b-2855de564de7" containerID="e46ecbca15b989fbf4b4d0b601955eb37da770af3bdb56b624aa574edd1b6d4e" exitCode=0 Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.369249 4809 generic.go:334] "Generic (PLEG): container finished" podID="c5182106-7776-4589-877b-2855de564de7" containerID="d1780c19b16939eeaaa8bc199ed0de29a951741ba1343c869efa054a59398bb7" exitCode=143 Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.369265 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5182106-7776-4589-877b-2855de564de7","Type":"ContainerDied","Data":"e46ecbca15b989fbf4b4d0b601955eb37da770af3bdb56b624aa574edd1b6d4e"} Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.369293 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5182106-7776-4589-877b-2855de564de7","Type":"ContainerDied","Data":"d1780c19b16939eeaaa8bc199ed0de29a951741ba1343c869efa054a59398bb7"} Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.369307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c5182106-7776-4589-877b-2855de564de7","Type":"ContainerDied","Data":"2ecfd02123498532083d30f016e18d7f17052084c3b7f3dc73028b8a8fb9a063"} Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.369319 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ecfd02123498532083d30f016e18d7f17052084c3b7f3dc73028b8a8fb9a063" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.377318 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3433b83d-e281-4415-9547-c1d2868331fa-kube-api-access-2k5pl" (OuterVolumeSpecName: "kube-api-access-2k5pl") pod "3433b83d-e281-4415-9547-c1d2868331fa" (UID: "3433b83d-e281-4415-9547-c1d2868331fa"). InnerVolumeSpecName "kube-api-access-2k5pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.378953 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-config-data" (OuterVolumeSpecName: "config-data") pod "3433b83d-e281-4415-9547-c1d2868331fa" (UID: "3433b83d-e281-4415-9547-c1d2868331fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.393907 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3433b83d-e281-4415-9547-c1d2868331fa" (UID: "3433b83d-e281-4415-9547-c1d2868331fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.401705 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:00 crc kubenswrapper[4809]: W1205 12:43:00.403941 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc097deb0_a548_4b13_a41f_68fa3b98074a.slice/crio-dd51f625902330dba56cde9760d25bed0bd328988c68e5f778162edd8176f1d1 WatchSource:0}: Error finding container dd51f625902330dba56cde9760d25bed0bd328988c68e5f778162edd8176f1d1: Status 404 returned error can't find the container with id dd51f625902330dba56cde9760d25bed0bd328988c68e5f778162edd8176f1d1 Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.424870 4809 scope.go:117] "RemoveContainer" containerID="1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.448792 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k5pl\" (UniqueName: \"kubernetes.io/projected/3433b83d-e281-4415-9547-c1d2868331fa-kube-api-access-2k5pl\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.448832 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3433b83d-e281-4415-9547-c1d2868331fa-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.448842 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.448851 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3433b83d-e281-4415-9547-c1d2868331fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.451800 4809 scope.go:117] "RemoveContainer" containerID="ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388" Dec 05 12:43:00 crc kubenswrapper[4809]: E1205 12:43:00.452244 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388\": container with ID starting with ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388 not found: ID does not exist" containerID="ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.452294 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388"} err="failed to get container status \"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388\": rpc error: code = NotFound desc = could not find container \"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388\": container with ID starting with ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388 not found: ID does not exist" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.452318 4809 scope.go:117] "RemoveContainer" containerID="1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389" Dec 05 12:43:00 crc kubenswrapper[4809]: E1205 12:43:00.453061 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389\": container with ID starting with 1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389 not found: ID does not exist" containerID="1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.453092 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389"} err="failed to get container status \"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389\": rpc error: code = NotFound desc = could not find container \"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389\": container with ID starting with 1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389 not found: ID does not exist" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.453115 4809 scope.go:117] "RemoveContainer" containerID="ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.453496 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388"} err="failed to get container status \"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388\": rpc error: code = NotFound desc = could not find container \"ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388\": container with ID starting with ce0b31ed200edd04c924e0914efc918407c6571eab22dc8b1c70c0d2dc3c3388 not found: ID does not exist" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.453524 4809 scope.go:117] "RemoveContainer" containerID="1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.453795 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389"} err="failed to get container status \"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389\": rpc error: code = NotFound desc = could not find container \"1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389\": container with ID starting with 1670b13f7dd6c797111db145ea46300c02f17c555e7db5a748084a012370e389 not found: ID does not exist" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.549880 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-config-data\") pod \"c5182106-7776-4589-877b-2855de564de7\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.550031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-combined-ca-bundle\") pod \"c5182106-7776-4589-877b-2855de564de7\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.550063 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h6x7\" (UniqueName: \"kubernetes.io/projected/c5182106-7776-4589-877b-2855de564de7-kube-api-access-6h6x7\") pod \"c5182106-7776-4589-877b-2855de564de7\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.550118 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5182106-7776-4589-877b-2855de564de7-logs\") pod \"c5182106-7776-4589-877b-2855de564de7\" (UID: \"c5182106-7776-4589-877b-2855de564de7\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.550874 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5182106-7776-4589-877b-2855de564de7-logs" (OuterVolumeSpecName: "logs") pod "c5182106-7776-4589-877b-2855de564de7" (UID: "c5182106-7776-4589-877b-2855de564de7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.554709 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5182106-7776-4589-877b-2855de564de7-kube-api-access-6h6x7" (OuterVolumeSpecName: "kube-api-access-6h6x7") pod "c5182106-7776-4589-877b-2855de564de7" (UID: "c5182106-7776-4589-877b-2855de564de7"). InnerVolumeSpecName "kube-api-access-6h6x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.587876 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-config-data" (OuterVolumeSpecName: "config-data") pod "c5182106-7776-4589-877b-2855de564de7" (UID: "c5182106-7776-4589-877b-2855de564de7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.590318 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5182106-7776-4589-877b-2855de564de7" (UID: "c5182106-7776-4589-877b-2855de564de7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.652438 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.653105 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5182106-7776-4589-877b-2855de564de7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.653123 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h6x7\" (UniqueName: \"kubernetes.io/projected/c5182106-7776-4589-877b-2855de564de7-kube-api-access-6h6x7\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.653135 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5182106-7776-4589-877b-2855de564de7-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.660608 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.783349 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.793092 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.806421 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:00 crc kubenswrapper[4809]: E1205 12:43:00.806877 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-log" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.806900 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-log" Dec 05 12:43:00 crc kubenswrapper[4809]: E1205 12:43:00.806929 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-metadata" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.806938 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-metadata" Dec 05 12:43:00 crc kubenswrapper[4809]: E1205 12:43:00.806955 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60eff64d-276b-41c5-ab5f-eda25d2a0c6c" containerName="nova-scheduler-scheduler" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.806964 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="60eff64d-276b-41c5-ab5f-eda25d2a0c6c" containerName="nova-scheduler-scheduler" Dec 05 12:43:00 crc kubenswrapper[4809]: E1205 12:43:00.806981 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-log" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.806989 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-log" Dec 05 12:43:00 crc kubenswrapper[4809]: E1205 12:43:00.807014 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-api" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.807022 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-api" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.807249 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-log" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.807278 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-log" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.807295 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3433b83d-e281-4415-9547-c1d2868331fa" containerName="nova-api-api" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.807306 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="60eff64d-276b-41c5-ab5f-eda25d2a0c6c" containerName="nova-scheduler-scheduler" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.807324 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5182106-7776-4589-877b-2855de564de7" containerName="nova-metadata-metadata" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.808444 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.811565 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.821810 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.857236 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqglw\" (UniqueName: \"kubernetes.io/projected/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-kube-api-access-mqglw\") pod \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.857415 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-config-data\") pod \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.857481 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-combined-ca-bundle\") pod \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\" (UID: \"60eff64d-276b-41c5-ab5f-eda25d2a0c6c\") " Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.876518 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-kube-api-access-mqglw" (OuterVolumeSpecName: "kube-api-access-mqglw") pod "60eff64d-276b-41c5-ab5f-eda25d2a0c6c" (UID: "60eff64d-276b-41c5-ab5f-eda25d2a0c6c"). InnerVolumeSpecName "kube-api-access-mqglw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.888814 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3433b83d-e281-4415-9547-c1d2868331fa" path="/var/lib/kubelet/pods/3433b83d-e281-4415-9547-c1d2868331fa/volumes" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.894715 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-config-data" (OuterVolumeSpecName: "config-data") pod "60eff64d-276b-41c5-ab5f-eda25d2a0c6c" (UID: "60eff64d-276b-41c5-ab5f-eda25d2a0c6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.905513 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60eff64d-276b-41c5-ab5f-eda25d2a0c6c" (UID: "60eff64d-276b-41c5-ab5f-eda25d2a0c6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.961739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-logs\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.961825 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.961862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4z2p\" (UniqueName: \"kubernetes.io/projected/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-kube-api-access-x4z2p\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.961888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-config-data\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.961994 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.962010 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:00 crc kubenswrapper[4809]: I1205 12:43:00.962024 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqglw\" (UniqueName: \"kubernetes.io/projected/60eff64d-276b-41c5-ab5f-eda25d2a0c6c-kube-api-access-mqglw\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.063461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-logs\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.063543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.063574 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4z2p\" (UniqueName: \"kubernetes.io/projected/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-kube-api-access-x4z2p\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.063601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-config-data\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.063942 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-logs\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.067556 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.067655 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-config-data\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.080717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4z2p\" (UniqueName: \"kubernetes.io/projected/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-kube-api-access-x4z2p\") pod \"nova-api-0\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.175058 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.381948 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.383083 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c097deb0-a548-4b13-a41f-68fa3b98074a","Type":"ContainerStarted","Data":"cfac4fb786d41315a7e45dd20833f28176477f22c645dc48ff0c421df86b5681"} Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.383121 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c097deb0-a548-4b13-a41f-68fa3b98074a","Type":"ContainerStarted","Data":"dd51f625902330dba56cde9760d25bed0bd328988c68e5f778162edd8176f1d1"} Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.383136 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.389571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"60eff64d-276b-41c5-ab5f-eda25d2a0c6c","Type":"ContainerDied","Data":"a3869bae9a5581f98b7c5a00ca0440ab0179d19ae922757f2c43a3beb748bb24"} Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.389648 4809 scope.go:117] "RemoveContainer" containerID="8cc99fa11e6dc7ae43979c4a85c0b43ba35752a89ca4360a07c64982202725d9" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.389658 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.389626 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.408369 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.408524 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.408495655 podStartE2EDuration="2.408495655s" podCreationTimestamp="2025-12-05 12:42:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:01.398394143 +0000 UTC m=+5676.789370691" watchObservedRunningTime="2025-12-05 12:43:01.408495655 +0000 UTC m=+5676.799472213" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.446050 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.465906 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.489710 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.513543 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.525981 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.528650 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.535165 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.539057 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.540942 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.544169 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.558977 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.582622 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.632439 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:01 crc kubenswrapper[4809]: W1205 12:43:01.637495 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6e4944f_9aa4_4c6d_b767_62f89b71d2a2.slice/crio-68c16d656b594695f7ec7cba42cb2a9fd691953f9d8368ac544a5784333be238 WatchSource:0}: Error finding container 68c16d656b594695f7ec7cba42cb2a9fd691953f9d8368ac544a5784333be238: Status 404 returned error can't find the container with id 68c16d656b594695f7ec7cba42cb2a9fd691953f9d8368ac544a5784333be238 Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.676896 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-config-data\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.676967 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.676991 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81179bfb-26f9-4ba4-a34a-d5b026de3727-logs\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.677032 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hxdc\" (UniqueName: \"kubernetes.io/projected/81179bfb-26f9-4ba4-a34a-d5b026de3727-kube-api-access-5hxdc\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.677083 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-config-data\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.677104 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.677147 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cfg2\" (UniqueName: \"kubernetes.io/projected/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-kube-api-access-2cfg2\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.778482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-config-data\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.778538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.778585 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81179bfb-26f9-4ba4-a34a-d5b026de3727-logs\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.778626 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hxdc\" (UniqueName: \"kubernetes.io/projected/81179bfb-26f9-4ba4-a34a-d5b026de3727-kube-api-access-5hxdc\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.778720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.778740 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-config-data\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.778779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cfg2\" (UniqueName: \"kubernetes.io/projected/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-kube-api-access-2cfg2\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.779839 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81179bfb-26f9-4ba4-a34a-d5b026de3727-logs\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.797678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.799415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-config-data\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.799622 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-config-data\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.799670 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.812488 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hxdc\" (UniqueName: \"kubernetes.io/projected/81179bfb-26f9-4ba4-a34a-d5b026de3727-kube-api-access-5hxdc\") pod \"nova-metadata-0\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.812519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cfg2\" (UniqueName: \"kubernetes.io/projected/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-kube-api-access-2cfg2\") pod \"nova-scheduler-0\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.827822 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.851440 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.868232 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.910835 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f894f5fcc-p99x8"] Dec 05 12:43:01 crc kubenswrapper[4809]: I1205 12:43:01.914584 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" podUID="c094fae2-a65d-4d07-9621-76b469e2776f" containerName="dnsmasq-dns" containerID="cri-o://689fbc286b1035d92ae104840f2203382ea47c3a6da66e170e529aa06b318a53" gracePeriod=10 Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.417111 4809 generic.go:334] "Generic (PLEG): container finished" podID="c094fae2-a65d-4d07-9621-76b469e2776f" containerID="689fbc286b1035d92ae104840f2203382ea47c3a6da66e170e529aa06b318a53" exitCode=0 Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.417490 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" event={"ID":"c094fae2-a65d-4d07-9621-76b469e2776f","Type":"ContainerDied","Data":"689fbc286b1035d92ae104840f2203382ea47c3a6da66e170e529aa06b318a53"} Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.423799 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2","Type":"ContainerStarted","Data":"5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114"} Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.423824 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2","Type":"ContainerStarted","Data":"e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8"} Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.423836 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2","Type":"ContainerStarted","Data":"68c16d656b594695f7ec7cba42cb2a9fd691953f9d8368ac544a5784333be238"} Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.447225 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.468390 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.4683602159999998 podStartE2EDuration="2.468360216s" podCreationTimestamp="2025-12-05 12:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:02.446887328 +0000 UTC m=+5677.837863886" watchObservedRunningTime="2025-12-05 12:43:02.468360216 +0000 UTC m=+5677.859336774" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.490586 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.527756 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.560480 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.701921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-sb\") pod \"c094fae2-a65d-4d07-9621-76b469e2776f\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.702466 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llcj7\" (UniqueName: \"kubernetes.io/projected/c094fae2-a65d-4d07-9621-76b469e2776f-kube-api-access-llcj7\") pod \"c094fae2-a65d-4d07-9621-76b469e2776f\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.702592 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-config\") pod \"c094fae2-a65d-4d07-9621-76b469e2776f\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.702666 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-dns-svc\") pod \"c094fae2-a65d-4d07-9621-76b469e2776f\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.702704 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-nb\") pod \"c094fae2-a65d-4d07-9621-76b469e2776f\" (UID: \"c094fae2-a65d-4d07-9621-76b469e2776f\") " Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.722216 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c094fae2-a65d-4d07-9621-76b469e2776f-kube-api-access-llcj7" (OuterVolumeSpecName: "kube-api-access-llcj7") pod "c094fae2-a65d-4d07-9621-76b469e2776f" (UID: "c094fae2-a65d-4d07-9621-76b469e2776f"). InnerVolumeSpecName "kube-api-access-llcj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.805581 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llcj7\" (UniqueName: \"kubernetes.io/projected/c094fae2-a65d-4d07-9621-76b469e2776f-kube-api-access-llcj7\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.834710 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c094fae2-a65d-4d07-9621-76b469e2776f" (UID: "c094fae2-a65d-4d07-9621-76b469e2776f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.841250 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c094fae2-a65d-4d07-9621-76b469e2776f" (UID: "c094fae2-a65d-4d07-9621-76b469e2776f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.846389 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c094fae2-a65d-4d07-9621-76b469e2776f" (UID: "c094fae2-a65d-4d07-9621-76b469e2776f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.856974 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-config" (OuterVolumeSpecName: "config") pod "c094fae2-a65d-4d07-9621-76b469e2776f" (UID: "c094fae2-a65d-4d07-9621-76b469e2776f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.896348 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60eff64d-276b-41c5-ab5f-eda25d2a0c6c" path="/var/lib/kubelet/pods/60eff64d-276b-41c5-ab5f-eda25d2a0c6c/volumes" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.897179 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5182106-7776-4589-877b-2855de564de7" path="/var/lib/kubelet/pods/c5182106-7776-4589-877b-2855de564de7/volumes" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.918716 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.918770 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.918783 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:02 crc kubenswrapper[4809]: I1205 12:43:02.918801 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c094fae2-a65d-4d07-9621-76b469e2776f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:03 crc kubenswrapper[4809]: E1205 12:43:03.125137 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc094fae2_a65d_4d07_9621_76b469e2776f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc094fae2_a65d_4d07_9621_76b469e2776f.slice/crio-4837a9523f9293da731ee3194e8e3fe845f96bf68ad69a9c52ee762cbcdcc01e\": RecentStats: unable to find data in memory cache]" Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.439627 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4f6967d-1d38-42a1-83de-5a3f220b5cb8","Type":"ContainerStarted","Data":"eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f"} Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.439695 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4f6967d-1d38-42a1-83de-5a3f220b5cb8","Type":"ContainerStarted","Data":"c4165bd85cdee484091f5d6dec2081b8ec3ef13e7387ad41a86cbbf81fbbedb8"} Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.443535 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.443697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f894f5fcc-p99x8" event={"ID":"c094fae2-a65d-4d07-9621-76b469e2776f","Type":"ContainerDied","Data":"4837a9523f9293da731ee3194e8e3fe845f96bf68ad69a9c52ee762cbcdcc01e"} Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.443852 4809 scope.go:117] "RemoveContainer" containerID="689fbc286b1035d92ae104840f2203382ea47c3a6da66e170e529aa06b318a53" Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.449764 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81179bfb-26f9-4ba4-a34a-d5b026de3727","Type":"ContainerStarted","Data":"86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e"} Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.449884 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81179bfb-26f9-4ba4-a34a-d5b026de3727","Type":"ContainerStarted","Data":"fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef"} Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.449899 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81179bfb-26f9-4ba4-a34a-d5b026de3727","Type":"ContainerStarted","Data":"de9bacdac8413fc7b6536ea3fe36b8131ae951fad29af5733a5447aab8397de1"} Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.465085 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.465050637 podStartE2EDuration="2.465050637s" podCreationTimestamp="2025-12-05 12:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:03.461255295 +0000 UTC m=+5678.852231873" watchObservedRunningTime="2025-12-05 12:43:03.465050637 +0000 UTC m=+5678.856027215" Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.488075 4809 scope.go:117] "RemoveContainer" containerID="b179cd74f936ab579f93f15048d2ed355f9714ed2528e179285f8c5fd64ec384" Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.491687 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.491666354 podStartE2EDuration="2.491666354s" podCreationTimestamp="2025-12-05 12:43:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:03.483878754 +0000 UTC m=+5678.874855322" watchObservedRunningTime="2025-12-05 12:43:03.491666354 +0000 UTC m=+5678.882642912" Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.511111 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f894f5fcc-p99x8"] Dec 05 12:43:03 crc kubenswrapper[4809]: I1205 12:43:03.519858 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f894f5fcc-p99x8"] Dec 05 12:43:04 crc kubenswrapper[4809]: I1205 12:43:04.891878 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c094fae2-a65d-4d07-9621-76b469e2776f" path="/var/lib/kubelet/pods/c094fae2-a65d-4d07-9621-76b469e2776f/volumes" Dec 05 12:43:06 crc kubenswrapper[4809]: I1205 12:43:06.852263 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:43:06 crc kubenswrapper[4809]: I1205 12:43:06.852861 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:43:06 crc kubenswrapper[4809]: I1205 12:43:06.869008 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 12:43:09 crc kubenswrapper[4809]: I1205 12:43:09.885163 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.350877 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9kvm2"] Dec 05 12:43:10 crc kubenswrapper[4809]: E1205 12:43:10.351636 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c094fae2-a65d-4d07-9621-76b469e2776f" containerName="init" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.351677 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c094fae2-a65d-4d07-9621-76b469e2776f" containerName="init" Dec 05 12:43:10 crc kubenswrapper[4809]: E1205 12:43:10.351696 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c094fae2-a65d-4d07-9621-76b469e2776f" containerName="dnsmasq-dns" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.351706 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c094fae2-a65d-4d07-9621-76b469e2776f" containerName="dnsmasq-dns" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.351916 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c094fae2-a65d-4d07-9621-76b469e2776f" containerName="dnsmasq-dns" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.352714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.355808 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.359655 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.363777 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9kvm2"] Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.465897 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tln2f\" (UniqueName: \"kubernetes.io/projected/366869f6-f555-4b3b-95fe-e274e7d86652-kube-api-access-tln2f\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.465949 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-scripts\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.466158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.466315 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-config-data\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.568719 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-config-data\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.568794 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tln2f\" (UniqueName: \"kubernetes.io/projected/366869f6-f555-4b3b-95fe-e274e7d86652-kube-api-access-tln2f\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.568819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-scripts\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.568866 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.575277 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-scripts\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.575313 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-config-data\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.575792 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.584067 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tln2f\" (UniqueName: \"kubernetes.io/projected/366869f6-f555-4b3b-95fe-e274e7d86652-kube-api-access-tln2f\") pod \"nova-cell1-cell-mapping-9kvm2\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:10 crc kubenswrapper[4809]: I1205 12:43:10.678501 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:11 crc kubenswrapper[4809]: W1205 12:43:11.155395 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod366869f6_f555_4b3b_95fe_e274e7d86652.slice/crio-383cf89122ed90eba5419d0023077543d0c8c65719508c44231e2384f32785b0 WatchSource:0}: Error finding container 383cf89122ed90eba5419d0023077543d0c8c65719508c44231e2384f32785b0: Status 404 returned error can't find the container with id 383cf89122ed90eba5419d0023077543d0c8c65719508c44231e2384f32785b0 Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.160820 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9kvm2"] Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.175481 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.175527 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.524809 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9kvm2" event={"ID":"366869f6-f555-4b3b-95fe-e274e7d86652","Type":"ContainerStarted","Data":"0c7adc9e5c2b4683aee4eee140d398f170c888a8277719398a288127039103c6"} Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.525158 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9kvm2" event={"ID":"366869f6-f555-4b3b-95fe-e274e7d86652","Type":"ContainerStarted","Data":"383cf89122ed90eba5419d0023077543d0c8c65719508c44231e2384f32785b0"} Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.542594 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9kvm2" podStartSLOduration=1.542571272 podStartE2EDuration="1.542571272s" podCreationTimestamp="2025-12-05 12:43:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:11.542339556 +0000 UTC m=+5686.933316114" watchObservedRunningTime="2025-12-05 12:43:11.542571272 +0000 UTC m=+5686.933547840" Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.852184 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.852231 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.870129 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 12:43:11 crc kubenswrapper[4809]: I1205 12:43:11.904085 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 12:43:12 crc kubenswrapper[4809]: I1205 12:43:12.257879 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.69:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:12 crc kubenswrapper[4809]: I1205 12:43:12.258190 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.69:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:12 crc kubenswrapper[4809]: I1205 12:43:12.650905 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 12:43:12 crc kubenswrapper[4809]: I1205 12:43:12.933843 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.70:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:12 crc kubenswrapper[4809]: I1205 12:43:12.933907 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.70:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.047261 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.047634 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.047709 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.048576 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.048680 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" gracePeriod=600 Dec 05 12:43:14 crc kubenswrapper[4809]: E1205 12:43:14.166013 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.551010 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" exitCode=0 Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.551304 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31"} Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.551336 4809 scope.go:117] "RemoveContainer" containerID="0d29fa2887bb281af96ca44548b0644914998505ad459688152d496f4d53f7af" Dec 05 12:43:14 crc kubenswrapper[4809]: I1205 12:43:14.552000 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:43:14 crc kubenswrapper[4809]: E1205 12:43:14.552233 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:43:16 crc kubenswrapper[4809]: I1205 12:43:16.580931 4809 generic.go:334] "Generic (PLEG): container finished" podID="366869f6-f555-4b3b-95fe-e274e7d86652" containerID="0c7adc9e5c2b4683aee4eee140d398f170c888a8277719398a288127039103c6" exitCode=0 Dec 05 12:43:16 crc kubenswrapper[4809]: I1205 12:43:16.580996 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9kvm2" event={"ID":"366869f6-f555-4b3b-95fe-e274e7d86652","Type":"ContainerDied","Data":"0c7adc9e5c2b4683aee4eee140d398f170c888a8277719398a288127039103c6"} Dec 05 12:43:17 crc kubenswrapper[4809]: I1205 12:43:17.919110 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.015401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-scripts\") pod \"366869f6-f555-4b3b-95fe-e274e7d86652\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.015519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-config-data\") pod \"366869f6-f555-4b3b-95fe-e274e7d86652\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.015555 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tln2f\" (UniqueName: \"kubernetes.io/projected/366869f6-f555-4b3b-95fe-e274e7d86652-kube-api-access-tln2f\") pod \"366869f6-f555-4b3b-95fe-e274e7d86652\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.015684 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-combined-ca-bundle\") pod \"366869f6-f555-4b3b-95fe-e274e7d86652\" (UID: \"366869f6-f555-4b3b-95fe-e274e7d86652\") " Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.020748 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366869f6-f555-4b3b-95fe-e274e7d86652-kube-api-access-tln2f" (OuterVolumeSpecName: "kube-api-access-tln2f") pod "366869f6-f555-4b3b-95fe-e274e7d86652" (UID: "366869f6-f555-4b3b-95fe-e274e7d86652"). InnerVolumeSpecName "kube-api-access-tln2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.023343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-scripts" (OuterVolumeSpecName: "scripts") pod "366869f6-f555-4b3b-95fe-e274e7d86652" (UID: "366869f6-f555-4b3b-95fe-e274e7d86652"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.039968 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-config-data" (OuterVolumeSpecName: "config-data") pod "366869f6-f555-4b3b-95fe-e274e7d86652" (UID: "366869f6-f555-4b3b-95fe-e274e7d86652"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.043795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "366869f6-f555-4b3b-95fe-e274e7d86652" (UID: "366869f6-f555-4b3b-95fe-e274e7d86652"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.118340 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.118993 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tln2f\" (UniqueName: \"kubernetes.io/projected/366869f6-f555-4b3b-95fe-e274e7d86652-kube-api-access-tln2f\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.119093 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.119149 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/366869f6-f555-4b3b-95fe-e274e7d86652-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.600735 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9kvm2" event={"ID":"366869f6-f555-4b3b-95fe-e274e7d86652","Type":"ContainerDied","Data":"383cf89122ed90eba5419d0023077543d0c8c65719508c44231e2384f32785b0"} Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.600787 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="383cf89122ed90eba5419d0023077543d0c8c65719508c44231e2384f32785b0" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.600813 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9kvm2" Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.790846 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.791456 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-log" containerID="cri-o://e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8" gracePeriod=30 Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.791998 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-api" containerID="cri-o://5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114" gracePeriod=30 Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.807193 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.807532 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f4f6967d-1d38-42a1-83de-5a3f220b5cb8" containerName="nova-scheduler-scheduler" containerID="cri-o://eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" gracePeriod=30 Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.841440 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.841753 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-log" containerID="cri-o://fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef" gracePeriod=30 Dec 05 12:43:18 crc kubenswrapper[4809]: I1205 12:43:18.841803 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-metadata" containerID="cri-o://86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e" gracePeriod=30 Dec 05 12:43:19 crc kubenswrapper[4809]: I1205 12:43:19.609755 4809 generic.go:334] "Generic (PLEG): container finished" podID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerID="fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef" exitCode=143 Dec 05 12:43:19 crc kubenswrapper[4809]: I1205 12:43:19.609813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81179bfb-26f9-4ba4-a34a-d5b026de3727","Type":"ContainerDied","Data":"fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef"} Dec 05 12:43:19 crc kubenswrapper[4809]: I1205 12:43:19.610937 4809 generic.go:334] "Generic (PLEG): container finished" podID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerID="e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8" exitCode=143 Dec 05 12:43:19 crc kubenswrapper[4809]: I1205 12:43:19.610962 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2","Type":"ContainerDied","Data":"e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8"} Dec 05 12:43:21 crc kubenswrapper[4809]: E1205 12:43:21.871087 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 12:43:21 crc kubenswrapper[4809]: E1205 12:43:21.872991 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 12:43:21 crc kubenswrapper[4809]: E1205 12:43:21.874769 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 12:43:21 crc kubenswrapper[4809]: E1205 12:43:21.874830 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f4f6967d-1d38-42a1-83de-5a3f220b5cb8" containerName="nova-scheduler-scheduler" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.472794 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.483690 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.593126 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-combined-ca-bundle\") pod \"81179bfb-26f9-4ba4-a34a-d5b026de3727\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.593206 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-config-data\") pod \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.593303 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81179bfb-26f9-4ba4-a34a-d5b026de3727-logs\") pod \"81179bfb-26f9-4ba4-a34a-d5b026de3727\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.593331 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hxdc\" (UniqueName: \"kubernetes.io/projected/81179bfb-26f9-4ba4-a34a-d5b026de3727-kube-api-access-5hxdc\") pod \"81179bfb-26f9-4ba4-a34a-d5b026de3727\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.593842 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81179bfb-26f9-4ba4-a34a-d5b026de3727-logs" (OuterVolumeSpecName: "logs") pod "81179bfb-26f9-4ba4-a34a-d5b026de3727" (UID: "81179bfb-26f9-4ba4-a34a-d5b026de3727"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.593971 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-combined-ca-bundle\") pod \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.594297 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-logs\") pod \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.594335 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-config-data\") pod \"81179bfb-26f9-4ba4-a34a-d5b026de3727\" (UID: \"81179bfb-26f9-4ba4-a34a-d5b026de3727\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.594377 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4z2p\" (UniqueName: \"kubernetes.io/projected/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-kube-api-access-x4z2p\") pod \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\" (UID: \"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2\") " Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.594851 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-logs" (OuterVolumeSpecName: "logs") pod "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" (UID: "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.595052 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.595071 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81179bfb-26f9-4ba4-a34a-d5b026de3727-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.598644 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-kube-api-access-x4z2p" (OuterVolumeSpecName: "kube-api-access-x4z2p") pod "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" (UID: "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2"). InnerVolumeSpecName "kube-api-access-x4z2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.598950 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81179bfb-26f9-4ba4-a34a-d5b026de3727-kube-api-access-5hxdc" (OuterVolumeSpecName: "kube-api-access-5hxdc") pod "81179bfb-26f9-4ba4-a34a-d5b026de3727" (UID: "81179bfb-26f9-4ba4-a34a-d5b026de3727"). InnerVolumeSpecName "kube-api-access-5hxdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.620493 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81179bfb-26f9-4ba4-a34a-d5b026de3727" (UID: "81179bfb-26f9-4ba4-a34a-d5b026de3727"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.621065 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-config-data" (OuterVolumeSpecName: "config-data") pod "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" (UID: "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.621196 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-config-data" (OuterVolumeSpecName: "config-data") pod "81179bfb-26f9-4ba4-a34a-d5b026de3727" (UID: "81179bfb-26f9-4ba4-a34a-d5b026de3727"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.624598 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" (UID: "d6e4944f-9aa4-4c6d-b767-62f89b71d2a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.639341 4809 generic.go:334] "Generic (PLEG): container finished" podID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerID="5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114" exitCode=0 Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.639440 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2","Type":"ContainerDied","Data":"5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114"} Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.639476 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6e4944f-9aa4-4c6d-b767-62f89b71d2a2","Type":"ContainerDied","Data":"68c16d656b594695f7ec7cba42cb2a9fd691953f9d8368ac544a5784333be238"} Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.639492 4809 scope.go:117] "RemoveContainer" containerID="5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.639821 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.642321 4809 generic.go:334] "Generic (PLEG): container finished" podID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerID="86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e" exitCode=0 Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.642433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81179bfb-26f9-4ba4-a34a-d5b026de3727","Type":"ContainerDied","Data":"86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e"} Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.642505 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81179bfb-26f9-4ba4-a34a-d5b026de3727","Type":"ContainerDied","Data":"de9bacdac8413fc7b6536ea3fe36b8131ae951fad29af5733a5447aab8397de1"} Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.642596 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.663326 4809 scope.go:117] "RemoveContainer" containerID="e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.682804 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.697952 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.697999 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.698019 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hxdc\" (UniqueName: \"kubernetes.io/projected/81179bfb-26f9-4ba4-a34a-d5b026de3727-kube-api-access-5hxdc\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.698038 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.698053 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81179bfb-26f9-4ba4-a34a-d5b026de3727-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.698069 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4z2p\" (UniqueName: \"kubernetes.io/projected/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2-kube-api-access-x4z2p\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.701677 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.708986 4809 scope.go:117] "RemoveContainer" containerID="5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114" Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.712270 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114\": container with ID starting with 5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114 not found: ID does not exist" containerID="5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.712325 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114"} err="failed to get container status \"5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114\": rpc error: code = NotFound desc = could not find container \"5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114\": container with ID starting with 5f708e594ad3708c691860de7bb413b8c9d5a05b62371a24feb6f42d33082114 not found: ID does not exist" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.712360 4809 scope.go:117] "RemoveContainer" containerID="e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8" Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.715276 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8\": container with ID starting with e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8 not found: ID does not exist" containerID="e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.715480 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8"} err="failed to get container status \"e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8\": rpc error: code = NotFound desc = could not find container \"e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8\": container with ID starting with e9e7b6b33061b6f26d908b7d18993725cda4370e54b2d48baa1eceadcea146c8 not found: ID does not exist" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.715585 4809 scope.go:117] "RemoveContainer" containerID="86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.718601 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.743547 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.747876 4809 scope.go:117] "RemoveContainer" containerID="fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.757508 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.757981 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-log" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.757998 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-log" Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.758015 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-api" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758021 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-api" Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.758029 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366869f6-f555-4b3b-95fe-e274e7d86652" containerName="nova-manage" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758035 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="366869f6-f555-4b3b-95fe-e274e7d86652" containerName="nova-manage" Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.758054 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-log" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758061 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-log" Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.758078 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-metadata" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758084 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-metadata" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758247 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-log" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758261 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-metadata" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758268 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" containerName="nova-api-api" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758281 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" containerName="nova-metadata-log" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.758291 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="366869f6-f555-4b3b-95fe-e274e7d86652" containerName="nova-manage" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.759258 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.769434 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.776253 4809 scope.go:117] "RemoveContainer" containerID="86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.776376 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.777505 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e\": container with ID starting with 86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e not found: ID does not exist" containerID="86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.777544 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e"} err="failed to get container status \"86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e\": rpc error: code = NotFound desc = could not find container \"86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e\": container with ID starting with 86689e0618579b054d1ff29e7870c8b9fe721fab5b912e171fae1c018f069e7e not found: ID does not exist" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.777568 4809 scope.go:117] "RemoveContainer" containerID="fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef" Dec 05 12:43:22 crc kubenswrapper[4809]: E1205 12:43:22.777881 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef\": container with ID starting with fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef not found: ID does not exist" containerID="fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.777908 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef"} err="failed to get container status \"fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef\": rpc error: code = NotFound desc = could not find container \"fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef\": container with ID starting with fc2d98e84d59a0a9587398c998060fa043331ca8635fb39997514196302ea9ef not found: ID does not exist" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.783060 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.785122 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.790348 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.799484 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.882437 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81179bfb-26f9-4ba4-a34a-d5b026de3727" path="/var/lib/kubelet/pods/81179bfb-26f9-4ba4-a34a-d5b026de3727/volumes" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.883089 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e4944f-9aa4-4c6d-b767-62f89b71d2a2" path="/var/lib/kubelet/pods/d6e4944f-9aa4-4c6d-b767-62f89b71d2a2/volumes" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.900974 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.901261 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-config-data\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.901310 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-config-data\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.901365 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-logs\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.901562 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.901611 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hckkn\" (UniqueName: \"kubernetes.io/projected/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-kube-api-access-hckkn\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.901721 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w9fg\" (UniqueName: \"kubernetes.io/projected/080276e3-3e86-4d71-a8fc-c5d4927e7031-kube-api-access-5w9fg\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:22 crc kubenswrapper[4809]: I1205 12:43:22.901804 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/080276e3-3e86-4d71-a8fc-c5d4927e7031-logs\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004094 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-logs\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004386 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hckkn\" (UniqueName: \"kubernetes.io/projected/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-kube-api-access-hckkn\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004488 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w9fg\" (UniqueName: \"kubernetes.io/projected/080276e3-3e86-4d71-a8fc-c5d4927e7031-kube-api-access-5w9fg\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/080276e3-3e86-4d71-a8fc-c5d4927e7031-logs\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004619 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004777 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-config-data\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.004816 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-config-data\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.008432 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/080276e3-3e86-4d71-a8fc-c5d4927e7031-logs\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.009473 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-logs\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.010362 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-config-data\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.010422 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.013347 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.017687 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-config-data\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.027287 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hckkn\" (UniqueName: \"kubernetes.io/projected/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-kube-api-access-hckkn\") pod \"nova-metadata-0\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.029480 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w9fg\" (UniqueName: \"kubernetes.io/projected/080276e3-3e86-4d71-a8fc-c5d4927e7031-kube-api-access-5w9fg\") pod \"nova-api-0\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.089768 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.105853 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.610564 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.682995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"080276e3-3e86-4d71-a8fc-c5d4927e7031","Type":"ContainerStarted","Data":"b245d98bafd84e724512feea83dff46fb358914281b3c1d5055d17d2effb0bdd"} Dec 05 12:43:23 crc kubenswrapper[4809]: I1205 12:43:23.704715 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:43:23 crc kubenswrapper[4809]: W1205 12:43:23.708718 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f8dc59a_423c_4df6_b0a5_bf1842d1a269.slice/crio-d5f8c0a77ddc747a14371da95677be32a728a90df2d633bc13eccc3c8e8474d4 WatchSource:0}: Error finding container d5f8c0a77ddc747a14371da95677be32a728a90df2d633bc13eccc3c8e8474d4: Status 404 returned error can't find the container with id d5f8c0a77ddc747a14371da95677be32a728a90df2d633bc13eccc3c8e8474d4 Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.569880 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.703978 4809 generic.go:334] "Generic (PLEG): container finished" podID="f4f6967d-1d38-42a1-83de-5a3f220b5cb8" containerID="eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" exitCode=0 Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.704134 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.704827 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4f6967d-1d38-42a1-83de-5a3f220b5cb8","Type":"ContainerDied","Data":"eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f"} Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.705001 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4f6967d-1d38-42a1-83de-5a3f220b5cb8","Type":"ContainerDied","Data":"c4165bd85cdee484091f5d6dec2081b8ec3ef13e7387ad41a86cbbf81fbbedb8"} Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.705033 4809 scope.go:117] "RemoveContainer" containerID="eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.708492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f8dc59a-423c-4df6-b0a5-bf1842d1a269","Type":"ContainerStarted","Data":"f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41"} Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.708536 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f8dc59a-423c-4df6-b0a5-bf1842d1a269","Type":"ContainerStarted","Data":"d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a"} Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.708550 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f8dc59a-423c-4df6-b0a5-bf1842d1a269","Type":"ContainerStarted","Data":"d5f8c0a77ddc747a14371da95677be32a728a90df2d633bc13eccc3c8e8474d4"} Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.714542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"080276e3-3e86-4d71-a8fc-c5d4927e7031","Type":"ContainerStarted","Data":"86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60"} Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.714591 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"080276e3-3e86-4d71-a8fc-c5d4927e7031","Type":"ContainerStarted","Data":"f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6"} Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.730348 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.730330194 podStartE2EDuration="2.730330194s" podCreationTimestamp="2025-12-05 12:43:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:24.729032959 +0000 UTC m=+5700.120009517" watchObservedRunningTime="2025-12-05 12:43:24.730330194 +0000 UTC m=+5700.121306752" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.744183 4809 scope.go:117] "RemoveContainer" containerID="eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.745678 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-combined-ca-bundle\") pod \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.745856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cfg2\" (UniqueName: \"kubernetes.io/projected/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-kube-api-access-2cfg2\") pod \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.745932 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-config-data\") pod \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\" (UID: \"f4f6967d-1d38-42a1-83de-5a3f220b5cb8\") " Dec 05 12:43:24 crc kubenswrapper[4809]: E1205 12:43:24.748130 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f\": container with ID starting with eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f not found: ID does not exist" containerID="eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.748185 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f"} err="failed to get container status \"eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f\": rpc error: code = NotFound desc = could not find container \"eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f\": container with ID starting with eff48ab65746115b0e4a3766a243d95c9f6b1b2e317c3d4e836fa2bb9c186f4f not found: ID does not exist" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.759748 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.759719735 podStartE2EDuration="2.759719735s" podCreationTimestamp="2025-12-05 12:43:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:24.745735368 +0000 UTC m=+5700.136711926" watchObservedRunningTime="2025-12-05 12:43:24.759719735 +0000 UTC m=+5700.150696293" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.768409 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-kube-api-access-2cfg2" (OuterVolumeSpecName: "kube-api-access-2cfg2") pod "f4f6967d-1d38-42a1-83de-5a3f220b5cb8" (UID: "f4f6967d-1d38-42a1-83de-5a3f220b5cb8"). InnerVolumeSpecName "kube-api-access-2cfg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.776307 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-config-data" (OuterVolumeSpecName: "config-data") pod "f4f6967d-1d38-42a1-83de-5a3f220b5cb8" (UID: "f4f6967d-1d38-42a1-83de-5a3f220b5cb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.780183 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4f6967d-1d38-42a1-83de-5a3f220b5cb8" (UID: "f4f6967d-1d38-42a1-83de-5a3f220b5cb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.848362 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.848662 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:24 crc kubenswrapper[4809]: I1205 12:43:24.848677 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cfg2\" (UniqueName: \"kubernetes.io/projected/f4f6967d-1d38-42a1-83de-5a3f220b5cb8-kube-api-access-2cfg2\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.060878 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.068688 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.079722 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:25 crc kubenswrapper[4809]: E1205 12:43:25.080759 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f6967d-1d38-42a1-83de-5a3f220b5cb8" containerName="nova-scheduler-scheduler" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.080794 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f6967d-1d38-42a1-83de-5a3f220b5cb8" containerName="nova-scheduler-scheduler" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.082433 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f6967d-1d38-42a1-83de-5a3f220b5cb8" containerName="nova-scheduler-scheduler" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.085090 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.088736 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.108217 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.254862 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdgs4\" (UniqueName: \"kubernetes.io/projected/408b563d-0ca7-48c0-82a3-83683536d42a-kube-api-access-kdgs4\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.254940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-config-data\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.254997 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.356857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.357023 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdgs4\" (UniqueName: \"kubernetes.io/projected/408b563d-0ca7-48c0-82a3-83683536d42a-kube-api-access-kdgs4\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.357069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-config-data\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.373920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.374793 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-config-data\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.375373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdgs4\" (UniqueName: \"kubernetes.io/projected/408b563d-0ca7-48c0-82a3-83683536d42a-kube-api-access-kdgs4\") pod \"nova-scheduler-0\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.422915 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:43:25 crc kubenswrapper[4809]: I1205 12:43:25.876952 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:43:25 crc kubenswrapper[4809]: W1205 12:43:25.886267 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod408b563d_0ca7_48c0_82a3_83683536d42a.slice/crio-e63a36c58725cc0e682b94283bf513aba2a8d6a004c3c2551c2e2e0b9ee34343 WatchSource:0}: Error finding container e63a36c58725cc0e682b94283bf513aba2a8d6a004c3c2551c2e2e0b9ee34343: Status 404 returned error can't find the container with id e63a36c58725cc0e682b94283bf513aba2a8d6a004c3c2551c2e2e0b9ee34343 Dec 05 12:43:26 crc kubenswrapper[4809]: I1205 12:43:26.732549 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"408b563d-0ca7-48c0-82a3-83683536d42a","Type":"ContainerStarted","Data":"0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5"} Dec 05 12:43:26 crc kubenswrapper[4809]: I1205 12:43:26.732857 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"408b563d-0ca7-48c0-82a3-83683536d42a","Type":"ContainerStarted","Data":"e63a36c58725cc0e682b94283bf513aba2a8d6a004c3c2551c2e2e0b9ee34343"} Dec 05 12:43:26 crc kubenswrapper[4809]: I1205 12:43:26.755305 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.7552877850000002 podStartE2EDuration="1.755287785s" podCreationTimestamp="2025-12-05 12:43:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:26.746183749 +0000 UTC m=+5702.137160307" watchObservedRunningTime="2025-12-05 12:43:26.755287785 +0000 UTC m=+5702.146264343" Dec 05 12:43:26 crc kubenswrapper[4809]: I1205 12:43:26.884312 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4f6967d-1d38-42a1-83de-5a3f220b5cb8" path="/var/lib/kubelet/pods/f4f6967d-1d38-42a1-83de-5a3f220b5cb8/volumes" Dec 05 12:43:27 crc kubenswrapper[4809]: I1205 12:43:27.871761 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:43:27 crc kubenswrapper[4809]: E1205 12:43:27.872377 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:43:28 crc kubenswrapper[4809]: I1205 12:43:28.109478 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:43:28 crc kubenswrapper[4809]: I1205 12:43:28.109535 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:43:30 crc kubenswrapper[4809]: I1205 12:43:30.424095 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 12:43:33 crc kubenswrapper[4809]: I1205 12:43:33.090686 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 12:43:33 crc kubenswrapper[4809]: I1205 12:43:33.091176 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 12:43:33 crc kubenswrapper[4809]: I1205 12:43:33.115527 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 12:43:33 crc kubenswrapper[4809]: I1205 12:43:33.116813 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 12:43:33 crc kubenswrapper[4809]: E1205 12:43:33.926818 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice/crio-c4165bd85cdee484091f5d6dec2081b8ec3ef13e7387ad41a86cbbf81fbbedb8\": RecentStats: unable to find data in memory cache]" Dec 05 12:43:34 crc kubenswrapper[4809]: I1205 12:43:34.130801 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:34 crc kubenswrapper[4809]: I1205 12:43:34.254052 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:34 crc kubenswrapper[4809]: I1205 12:43:34.255003 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:34 crc kubenswrapper[4809]: I1205 12:43:34.255727 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:43:35 crc kubenswrapper[4809]: I1205 12:43:35.063421 4809 scope.go:117] "RemoveContainer" containerID="fb80e42d7da1e672dc0782f1c78f7440ac65eb1b2f9f612d4c2ce0d165acd6c7" Dec 05 12:43:35 crc kubenswrapper[4809]: I1205 12:43:35.424082 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 12:43:35 crc kubenswrapper[4809]: I1205 12:43:35.448598 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 12:43:35 crc kubenswrapper[4809]: I1205 12:43:35.844908 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 12:43:39 crc kubenswrapper[4809]: I1205 12:43:39.872304 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:43:39 crc kubenswrapper[4809]: E1205 12:43:39.873109 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.094790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.095240 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.095670 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.095718 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.098900 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.099514 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.111515 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.115583 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.118777 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.312005 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-858bd85b85-6f4h2"] Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.313906 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.325347 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-858bd85b85-6f4h2"] Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.404733 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-nb\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.404816 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-config\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.404893 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-sb\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.404934 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-dns-svc\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.404989 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p46wt\" (UniqueName: \"kubernetes.io/projected/07125729-91e5-46ca-a996-a8c24cd60ab8-kube-api-access-p46wt\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.506154 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-sb\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.506220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-dns-svc\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.506267 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p46wt\" (UniqueName: \"kubernetes.io/projected/07125729-91e5-46ca-a996-a8c24cd60ab8-kube-api-access-p46wt\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.506336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-nb\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.506365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-config\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.507106 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-dns-svc\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.507140 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-config\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.507290 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-nb\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.507649 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-sb\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.527844 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p46wt\" (UniqueName: \"kubernetes.io/projected/07125729-91e5-46ca-a996-a8c24cd60ab8-kube-api-access-p46wt\") pod \"dnsmasq-dns-858bd85b85-6f4h2\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.648349 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:43 crc kubenswrapper[4809]: I1205 12:43:43.939060 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 12:43:44 crc kubenswrapper[4809]: E1205 12:43:44.171943 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice/crio-c4165bd85cdee484091f5d6dec2081b8ec3ef13e7387ad41a86cbbf81fbbedb8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice\": RecentStats: unable to find data in memory cache]" Dec 05 12:43:44 crc kubenswrapper[4809]: I1205 12:43:44.185105 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-858bd85b85-6f4h2"] Dec 05 12:43:44 crc kubenswrapper[4809]: I1205 12:43:44.908405 4809 generic.go:334] "Generic (PLEG): container finished" podID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerID="1f24312485989194c8523337b87143359da0d902a5d7d4601ce2b4e43cb92953" exitCode=0 Dec 05 12:43:44 crc kubenswrapper[4809]: I1205 12:43:44.909939 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" event={"ID":"07125729-91e5-46ca-a996-a8c24cd60ab8","Type":"ContainerDied","Data":"1f24312485989194c8523337b87143359da0d902a5d7d4601ce2b4e43cb92953"} Dec 05 12:43:44 crc kubenswrapper[4809]: I1205 12:43:44.909973 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" event={"ID":"07125729-91e5-46ca-a996-a8c24cd60ab8","Type":"ContainerStarted","Data":"146014f3a9ad620720fbabad77325a596414010cf90155c41977e48d0b3cff08"} Dec 05 12:43:45 crc kubenswrapper[4809]: I1205 12:43:45.918944 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" event={"ID":"07125729-91e5-46ca-a996-a8c24cd60ab8","Type":"ContainerStarted","Data":"77dd2145995dbf3d1876b48cc6df8ba035b66367f40ce343ee99670367794002"} Dec 05 12:43:45 crc kubenswrapper[4809]: I1205 12:43:45.919287 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:45 crc kubenswrapper[4809]: I1205 12:43:45.944361 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" podStartSLOduration=2.94433904 podStartE2EDuration="2.94433904s" podCreationTimestamp="2025-12-05 12:43:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:43:45.935114562 +0000 UTC m=+5721.326091140" watchObservedRunningTime="2025-12-05 12:43:45.94433904 +0000 UTC m=+5721.335315598" Dec 05 12:43:51 crc kubenswrapper[4809]: I1205 12:43:51.872248 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:43:51 crc kubenswrapper[4809]: E1205 12:43:51.873531 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:43:53 crc kubenswrapper[4809]: I1205 12:43:53.650847 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:43:53 crc kubenswrapper[4809]: I1205 12:43:53.742373 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd474b5b7-sm5l9"] Dec 05 12:43:53 crc kubenswrapper[4809]: I1205 12:43:53.742639 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" podUID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerName="dnsmasq-dns" containerID="cri-o://7be44412e11c190acedf5a4d6d50042b56c1a3311699c368a608ccea05ed1abe" gracePeriod=10 Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.021532 4809 generic.go:334] "Generic (PLEG): container finished" podID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerID="7be44412e11c190acedf5a4d6d50042b56c1a3311699c368a608ccea05ed1abe" exitCode=0 Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.021860 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" event={"ID":"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1","Type":"ContainerDied","Data":"7be44412e11c190acedf5a4d6d50042b56c1a3311699c368a608ccea05ed1abe"} Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.284872 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.436483 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-sb\") pod \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.437053 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpgfg\" (UniqueName: \"kubernetes.io/projected/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-kube-api-access-vpgfg\") pod \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.437092 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-nb\") pod \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.437223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-config\") pod \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.437322 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-dns-svc\") pod \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\" (UID: \"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1\") " Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.442106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-kube-api-access-vpgfg" (OuterVolumeSpecName: "kube-api-access-vpgfg") pod "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" (UID: "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1"). InnerVolumeSpecName "kube-api-access-vpgfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:54 crc kubenswrapper[4809]: E1205 12:43:54.467411 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice/crio-c4165bd85cdee484091f5d6dec2081b8ec3ef13e7387ad41a86cbbf81fbbedb8\": RecentStats: unable to find data in memory cache]" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.483333 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-config" (OuterVolumeSpecName: "config") pod "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" (UID: "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.485135 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" (UID: "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.493398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" (UID: "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.509690 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" (UID: "56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.540747 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.540788 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.540799 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.540809 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpgfg\" (UniqueName: \"kubernetes.io/projected/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-kube-api-access-vpgfg\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:54 crc kubenswrapper[4809]: I1205 12:43:54.540819 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.043079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" event={"ID":"56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1","Type":"ContainerDied","Data":"d76b32bdd6c817a87f167dce0b662c529f2b0efe8f7bd6465e9481064172e860"} Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.043152 4809 scope.go:117] "RemoveContainer" containerID="7be44412e11c190acedf5a4d6d50042b56c1a3311699c368a608ccea05ed1abe" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.043366 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd474b5b7-sm5l9" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.069200 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd474b5b7-sm5l9"] Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.076544 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bd474b5b7-sm5l9"] Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.076772 4809 scope.go:117] "RemoveContainer" containerID="e96bafdbb4545cd249866567632a0a6220682b58d7e7c3aec925645c302b8ea3" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.408350 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wj5vp"] Dec 05 12:43:55 crc kubenswrapper[4809]: E1205 12:43:55.409006 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerName="init" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.409022 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerName="init" Dec 05 12:43:55 crc kubenswrapper[4809]: E1205 12:43:55.409050 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerName="dnsmasq-dns" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.409056 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerName="dnsmasq-dns" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.409230 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" containerName="dnsmasq-dns" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.409863 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.422836 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wj5vp"] Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.513679 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7e82-account-create-update-xwx56"] Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.515817 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.519307 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.524501 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7e82-account-create-update-xwx56"] Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.583164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-operator-scripts\") pod \"cinder-db-create-wj5vp\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.583370 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttw2c\" (UniqueName: \"kubernetes.io/projected/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-kube-api-access-ttw2c\") pod \"cinder-db-create-wj5vp\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.685103 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-operator-scripts\") pod \"cinder-7e82-account-create-update-xwx56\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.685194 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-operator-scripts\") pod \"cinder-db-create-wj5vp\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.685275 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrfq8\" (UniqueName: \"kubernetes.io/projected/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-kube-api-access-qrfq8\") pod \"cinder-7e82-account-create-update-xwx56\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.685350 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttw2c\" (UniqueName: \"kubernetes.io/projected/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-kube-api-access-ttw2c\") pod \"cinder-db-create-wj5vp\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.688310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-operator-scripts\") pod \"cinder-db-create-wj5vp\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.704957 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttw2c\" (UniqueName: \"kubernetes.io/projected/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-kube-api-access-ttw2c\") pod \"cinder-db-create-wj5vp\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.726927 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.787268 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrfq8\" (UniqueName: \"kubernetes.io/projected/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-kube-api-access-qrfq8\") pod \"cinder-7e82-account-create-update-xwx56\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.787414 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-operator-scripts\") pod \"cinder-7e82-account-create-update-xwx56\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.788337 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-operator-scripts\") pod \"cinder-7e82-account-create-update-xwx56\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.806458 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrfq8\" (UniqueName: \"kubernetes.io/projected/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-kube-api-access-qrfq8\") pod \"cinder-7e82-account-create-update-xwx56\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:55 crc kubenswrapper[4809]: I1205 12:43:55.834987 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:56 crc kubenswrapper[4809]: I1205 12:43:56.200393 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wj5vp"] Dec 05 12:43:56 crc kubenswrapper[4809]: I1205 12:43:56.313698 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7e82-account-create-update-xwx56"] Dec 05 12:43:56 crc kubenswrapper[4809]: W1205 12:43:56.318416 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0f9da55_e8ce_4c11_acd6_b3193cdb660d.slice/crio-cbd50b12418c182aaaecf7c4ad29baafe09d1c3cb26c61e6eaf2e16acf6fe9f2 WatchSource:0}: Error finding container cbd50b12418c182aaaecf7c4ad29baafe09d1c3cb26c61e6eaf2e16acf6fe9f2: Status 404 returned error can't find the container with id cbd50b12418c182aaaecf7c4ad29baafe09d1c3cb26c61e6eaf2e16acf6fe9f2 Dec 05 12:43:56 crc kubenswrapper[4809]: I1205 12:43:56.882814 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1" path="/var/lib/kubelet/pods/56fb7d0e-51fd-42c6-a4e4-2d8953d3a4d1/volumes" Dec 05 12:43:57 crc kubenswrapper[4809]: I1205 12:43:57.072272 4809 generic.go:334] "Generic (PLEG): container finished" podID="d574f3bf-cb38-4fbb-bb18-0f3135ee2630" containerID="9d3bea26c5302a972868cf3ac3d659712c4eaf2bac42f875ceec85c1a391a9be" exitCode=0 Dec 05 12:43:57 crc kubenswrapper[4809]: I1205 12:43:57.072347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wj5vp" event={"ID":"d574f3bf-cb38-4fbb-bb18-0f3135ee2630","Type":"ContainerDied","Data":"9d3bea26c5302a972868cf3ac3d659712c4eaf2bac42f875ceec85c1a391a9be"} Dec 05 12:43:57 crc kubenswrapper[4809]: I1205 12:43:57.072375 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wj5vp" event={"ID":"d574f3bf-cb38-4fbb-bb18-0f3135ee2630","Type":"ContainerStarted","Data":"1f20c63366ae694a06c6f6f8bb9dfd3d427d7a66286e012f75f366f7f0dcd9fc"} Dec 05 12:43:57 crc kubenswrapper[4809]: I1205 12:43:57.073736 4809 generic.go:334] "Generic (PLEG): container finished" podID="c0f9da55-e8ce-4c11-acd6-b3193cdb660d" containerID="13b13f2d8be871d2d4e0ba09737ccbfa5805a70eb3c088f9e0111d65fd96ba6d" exitCode=0 Dec 05 12:43:57 crc kubenswrapper[4809]: I1205 12:43:57.073765 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7e82-account-create-update-xwx56" event={"ID":"c0f9da55-e8ce-4c11-acd6-b3193cdb660d","Type":"ContainerDied","Data":"13b13f2d8be871d2d4e0ba09737ccbfa5805a70eb3c088f9e0111d65fd96ba6d"} Dec 05 12:43:57 crc kubenswrapper[4809]: I1205 12:43:57.073778 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7e82-account-create-update-xwx56" event={"ID":"c0f9da55-e8ce-4c11-acd6-b3193cdb660d","Type":"ContainerStarted","Data":"cbd50b12418c182aaaecf7c4ad29baafe09d1c3cb26c61e6eaf2e16acf6fe9f2"} Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.540579 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.551675 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.640825 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-operator-scripts\") pod \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.640969 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrfq8\" (UniqueName: \"kubernetes.io/projected/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-kube-api-access-qrfq8\") pod \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\" (UID: \"c0f9da55-e8ce-4c11-acd6-b3193cdb660d\") " Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.641070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttw2c\" (UniqueName: \"kubernetes.io/projected/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-kube-api-access-ttw2c\") pod \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.641125 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-operator-scripts\") pod \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\" (UID: \"d574f3bf-cb38-4fbb-bb18-0f3135ee2630\") " Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.641506 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c0f9da55-e8ce-4c11-acd6-b3193cdb660d" (UID: "c0f9da55-e8ce-4c11-acd6-b3193cdb660d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.641620 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d574f3bf-cb38-4fbb-bb18-0f3135ee2630" (UID: "d574f3bf-cb38-4fbb-bb18-0f3135ee2630"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.641680 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.646450 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-kube-api-access-ttw2c" (OuterVolumeSpecName: "kube-api-access-ttw2c") pod "d574f3bf-cb38-4fbb-bb18-0f3135ee2630" (UID: "d574f3bf-cb38-4fbb-bb18-0f3135ee2630"). InnerVolumeSpecName "kube-api-access-ttw2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.653767 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-kube-api-access-qrfq8" (OuterVolumeSpecName: "kube-api-access-qrfq8") pod "c0f9da55-e8ce-4c11-acd6-b3193cdb660d" (UID: "c0f9da55-e8ce-4c11-acd6-b3193cdb660d"). InnerVolumeSpecName "kube-api-access-qrfq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.743265 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrfq8\" (UniqueName: \"kubernetes.io/projected/c0f9da55-e8ce-4c11-acd6-b3193cdb660d-kube-api-access-qrfq8\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.743298 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttw2c\" (UniqueName: \"kubernetes.io/projected/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-kube-api-access-ttw2c\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:58 crc kubenswrapper[4809]: I1205 12:43:58.743309 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d574f3bf-cb38-4fbb-bb18-0f3135ee2630-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:43:59 crc kubenswrapper[4809]: I1205 12:43:59.091828 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wj5vp" event={"ID":"d574f3bf-cb38-4fbb-bb18-0f3135ee2630","Type":"ContainerDied","Data":"1f20c63366ae694a06c6f6f8bb9dfd3d427d7a66286e012f75f366f7f0dcd9fc"} Dec 05 12:43:59 crc kubenswrapper[4809]: I1205 12:43:59.092200 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f20c63366ae694a06c6f6f8bb9dfd3d427d7a66286e012f75f366f7f0dcd9fc" Dec 05 12:43:59 crc kubenswrapper[4809]: I1205 12:43:59.091862 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wj5vp" Dec 05 12:43:59 crc kubenswrapper[4809]: I1205 12:43:59.093255 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7e82-account-create-update-xwx56" event={"ID":"c0f9da55-e8ce-4c11-acd6-b3193cdb660d","Type":"ContainerDied","Data":"cbd50b12418c182aaaecf7c4ad29baafe09d1c3cb26c61e6eaf2e16acf6fe9f2"} Dec 05 12:43:59 crc kubenswrapper[4809]: I1205 12:43:59.093277 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbd50b12418c182aaaecf7c4ad29baafe09d1c3cb26c61e6eaf2e16acf6fe9f2" Dec 05 12:43:59 crc kubenswrapper[4809]: I1205 12:43:59.093291 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7e82-account-create-update-xwx56" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.713524 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-nr4kd"] Dec 05 12:44:00 crc kubenswrapper[4809]: E1205 12:44:00.714400 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d574f3bf-cb38-4fbb-bb18-0f3135ee2630" containerName="mariadb-database-create" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.714423 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d574f3bf-cb38-4fbb-bb18-0f3135ee2630" containerName="mariadb-database-create" Dec 05 12:44:00 crc kubenswrapper[4809]: E1205 12:44:00.714449 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f9da55-e8ce-4c11-acd6-b3193cdb660d" containerName="mariadb-account-create-update" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.714459 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f9da55-e8ce-4c11-acd6-b3193cdb660d" containerName="mariadb-account-create-update" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.714743 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d574f3bf-cb38-4fbb-bb18-0f3135ee2630" containerName="mariadb-database-create" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.714774 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0f9da55-e8ce-4c11-acd6-b3193cdb660d" containerName="mariadb-account-create-update" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.715609 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.720170 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.720408 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-78drc" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.721031 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.748778 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nr4kd"] Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.883452 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-combined-ca-bundle\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.883509 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a08109f6-e55a-4a76-a817-d788341d85b0-etc-machine-id\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.883533 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-config-data\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.883653 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-scripts\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.883830 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpsw9\" (UniqueName: \"kubernetes.io/projected/a08109f6-e55a-4a76-a817-d788341d85b0-kube-api-access-mpsw9\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.883935 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-db-sync-config-data\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.985449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-combined-ca-bundle\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.985510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a08109f6-e55a-4a76-a817-d788341d85b0-etc-machine-id\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.985550 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-config-data\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.985573 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-scripts\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.985643 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpsw9\" (UniqueName: \"kubernetes.io/projected/a08109f6-e55a-4a76-a817-d788341d85b0-kube-api-access-mpsw9\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.985680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-db-sync-config-data\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.985680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a08109f6-e55a-4a76-a817-d788341d85b0-etc-machine-id\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.991565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-scripts\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.993689 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-db-sync-config-data\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.994255 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-combined-ca-bundle\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:00 crc kubenswrapper[4809]: I1205 12:44:00.998444 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-config-data\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:01 crc kubenswrapper[4809]: I1205 12:44:01.003530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpsw9\" (UniqueName: \"kubernetes.io/projected/a08109f6-e55a-4a76-a817-d788341d85b0-kube-api-access-mpsw9\") pod \"cinder-db-sync-nr4kd\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:01 crc kubenswrapper[4809]: I1205 12:44:01.049131 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:01 crc kubenswrapper[4809]: I1205 12:44:01.504809 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nr4kd"] Dec 05 12:44:02 crc kubenswrapper[4809]: I1205 12:44:02.125440 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nr4kd" event={"ID":"a08109f6-e55a-4a76-a817-d788341d85b0","Type":"ContainerStarted","Data":"3e8899852190025ee24e93eadf6e4616ad17b3ee16374f41bba0681481766038"} Dec 05 12:44:03 crc kubenswrapper[4809]: I1205 12:44:03.161606 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nr4kd" event={"ID":"a08109f6-e55a-4a76-a817-d788341d85b0","Type":"ContainerStarted","Data":"949f49a3229508b2640e738584bb8fe8a3049ab23a7427a6f62ddeaae00ccc77"} Dec 05 12:44:03 crc kubenswrapper[4809]: I1205 12:44:03.186673 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-nr4kd" podStartSLOduration=3.186615709 podStartE2EDuration="3.186615709s" podCreationTimestamp="2025-12-05 12:44:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:03.176729883 +0000 UTC m=+5738.567706471" watchObservedRunningTime="2025-12-05 12:44:03.186615709 +0000 UTC m=+5738.577592267" Dec 05 12:44:04 crc kubenswrapper[4809]: E1205 12:44:04.696198 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice/crio-c4165bd85cdee484091f5d6dec2081b8ec3ef13e7387ad41a86cbbf81fbbedb8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice\": RecentStats: unable to find data in memory cache]" Dec 05 12:44:04 crc kubenswrapper[4809]: I1205 12:44:04.879167 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:44:04 crc kubenswrapper[4809]: E1205 12:44:04.879898 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:44:05 crc kubenswrapper[4809]: I1205 12:44:05.184866 4809 generic.go:334] "Generic (PLEG): container finished" podID="a08109f6-e55a-4a76-a817-d788341d85b0" containerID="949f49a3229508b2640e738584bb8fe8a3049ab23a7427a6f62ddeaae00ccc77" exitCode=0 Dec 05 12:44:05 crc kubenswrapper[4809]: I1205 12:44:05.184920 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nr4kd" event={"ID":"a08109f6-e55a-4a76-a817-d788341d85b0","Type":"ContainerDied","Data":"949f49a3229508b2640e738584bb8fe8a3049ab23a7427a6f62ddeaae00ccc77"} Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.541296 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.703152 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-combined-ca-bundle\") pod \"a08109f6-e55a-4a76-a817-d788341d85b0\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.703500 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpsw9\" (UniqueName: \"kubernetes.io/projected/a08109f6-e55a-4a76-a817-d788341d85b0-kube-api-access-mpsw9\") pod \"a08109f6-e55a-4a76-a817-d788341d85b0\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.703729 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-scripts\") pod \"a08109f6-e55a-4a76-a817-d788341d85b0\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.703879 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-config-data\") pod \"a08109f6-e55a-4a76-a817-d788341d85b0\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.703993 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-db-sync-config-data\") pod \"a08109f6-e55a-4a76-a817-d788341d85b0\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.704119 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a08109f6-e55a-4a76-a817-d788341d85b0-etc-machine-id\") pod \"a08109f6-e55a-4a76-a817-d788341d85b0\" (UID: \"a08109f6-e55a-4a76-a817-d788341d85b0\") " Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.704266 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a08109f6-e55a-4a76-a817-d788341d85b0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a08109f6-e55a-4a76-a817-d788341d85b0" (UID: "a08109f6-e55a-4a76-a817-d788341d85b0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.704830 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a08109f6-e55a-4a76-a817-d788341d85b0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.711824 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-scripts" (OuterVolumeSpecName: "scripts") pod "a08109f6-e55a-4a76-a817-d788341d85b0" (UID: "a08109f6-e55a-4a76-a817-d788341d85b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.711889 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a08109f6-e55a-4a76-a817-d788341d85b0" (UID: "a08109f6-e55a-4a76-a817-d788341d85b0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.711994 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a08109f6-e55a-4a76-a817-d788341d85b0-kube-api-access-mpsw9" (OuterVolumeSpecName: "kube-api-access-mpsw9") pod "a08109f6-e55a-4a76-a817-d788341d85b0" (UID: "a08109f6-e55a-4a76-a817-d788341d85b0"). InnerVolumeSpecName "kube-api-access-mpsw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.730056 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a08109f6-e55a-4a76-a817-d788341d85b0" (UID: "a08109f6-e55a-4a76-a817-d788341d85b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.748868 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-config-data" (OuterVolumeSpecName: "config-data") pod "a08109f6-e55a-4a76-a817-d788341d85b0" (UID: "a08109f6-e55a-4a76-a817-d788341d85b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.806923 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.806959 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpsw9\" (UniqueName: \"kubernetes.io/projected/a08109f6-e55a-4a76-a817-d788341d85b0-kube-api-access-mpsw9\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.806971 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.806980 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:06 crc kubenswrapper[4809]: I1205 12:44:06.806988 4809 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a08109f6-e55a-4a76-a817-d788341d85b0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.207182 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nr4kd" event={"ID":"a08109f6-e55a-4a76-a817-d788341d85b0","Type":"ContainerDied","Data":"3e8899852190025ee24e93eadf6e4616ad17b3ee16374f41bba0681481766038"} Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.207685 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e8899852190025ee24e93eadf6e4616ad17b3ee16374f41bba0681481766038" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.207288 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nr4kd" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.491789 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6974cc94df-frb66"] Dec 05 12:44:07 crc kubenswrapper[4809]: E1205 12:44:07.492197 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a08109f6-e55a-4a76-a817-d788341d85b0" containerName="cinder-db-sync" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.492233 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a08109f6-e55a-4a76-a817-d788341d85b0" containerName="cinder-db-sync" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.492437 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a08109f6-e55a-4a76-a817-d788341d85b0" containerName="cinder-db-sync" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.493435 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.521325 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6974cc94df-frb66"] Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.620837 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-sb\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.620932 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrkqv\" (UniqueName: \"kubernetes.io/projected/9da25051-70b9-4014-b10c-900022a35a1a-kube-api-access-vrkqv\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.621072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-dns-svc\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.621119 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-nb\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.621288 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-config\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.651210 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.653262 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.655855 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.655969 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.658111 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.658913 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-78drc" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.665697 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.722938 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrkqv\" (UniqueName: \"kubernetes.io/projected/9da25051-70b9-4014-b10c-900022a35a1a-kube-api-access-vrkqv\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.723037 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-dns-svc\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.723071 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-nb\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.723116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-config\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.723155 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-sb\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.724036 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-sb\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.724560 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-dns-svc\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.724688 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-config\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.724848 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-nb\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.742448 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrkqv\" (UniqueName: \"kubernetes.io/projected/9da25051-70b9-4014-b10c-900022a35a1a-kube-api-access-vrkqv\") pod \"dnsmasq-dns-6974cc94df-frb66\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.825234 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhvqv\" (UniqueName: \"kubernetes.io/projected/38ef0e35-a057-43e9-901d-e02be6abcca7-kube-api-access-xhvqv\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.825304 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ef0e35-a057-43e9-901d-e02be6abcca7-logs\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.825340 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.825478 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.825638 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data-custom\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.825702 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38ef0e35-a057-43e9-901d-e02be6abcca7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.825732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-scripts\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.826129 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.931527 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.931920 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhvqv\" (UniqueName: \"kubernetes.io/projected/38ef0e35-a057-43e9-901d-e02be6abcca7-kube-api-access-xhvqv\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.931961 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ef0e35-a057-43e9-901d-e02be6abcca7-logs\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.931986 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.932014 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data-custom\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.932052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38ef0e35-a057-43e9-901d-e02be6abcca7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.932069 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-scripts\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.933203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ef0e35-a057-43e9-901d-e02be6abcca7-logs\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.935410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38ef0e35-a057-43e9-901d-e02be6abcca7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.937406 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.940237 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data-custom\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.954201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.944623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-scripts\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.967006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhvqv\" (UniqueName: \"kubernetes.io/projected/38ef0e35-a057-43e9-901d-e02be6abcca7-kube-api-access-xhvqv\") pod \"cinder-api-0\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " pod="openstack/cinder-api-0" Dec 05 12:44:07 crc kubenswrapper[4809]: I1205 12:44:07.981260 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 12:44:08 crc kubenswrapper[4809]: I1205 12:44:08.369567 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6974cc94df-frb66"] Dec 05 12:44:08 crc kubenswrapper[4809]: I1205 12:44:08.530272 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:08 crc kubenswrapper[4809]: W1205 12:44:08.534666 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38ef0e35_a057_43e9_901d_e02be6abcca7.slice/crio-0b2a3a8debd7335a0ff5356a67e44a5261e15e8aded8fa0cee8d9d344f76929f WatchSource:0}: Error finding container 0b2a3a8debd7335a0ff5356a67e44a5261e15e8aded8fa0cee8d9d344f76929f: Status 404 returned error can't find the container with id 0b2a3a8debd7335a0ff5356a67e44a5261e15e8aded8fa0cee8d9d344f76929f Dec 05 12:44:09 crc kubenswrapper[4809]: I1205 12:44:09.240679 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"38ef0e35-a057-43e9-901d-e02be6abcca7","Type":"ContainerStarted","Data":"fbd933c23104016a5dbdbd261a9c6c40593e4cf9ec15ee5cdeaed66e05ae6a7e"} Dec 05 12:44:09 crc kubenswrapper[4809]: I1205 12:44:09.241006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"38ef0e35-a057-43e9-901d-e02be6abcca7","Type":"ContainerStarted","Data":"0b2a3a8debd7335a0ff5356a67e44a5261e15e8aded8fa0cee8d9d344f76929f"} Dec 05 12:44:09 crc kubenswrapper[4809]: I1205 12:44:09.242561 4809 generic.go:334] "Generic (PLEG): container finished" podID="9da25051-70b9-4014-b10c-900022a35a1a" containerID="0c311f82c9edb7b608b237a67d4bd4e751e5f1ce3fd4a92d8f4043b66d73021e" exitCode=0 Dec 05 12:44:09 crc kubenswrapper[4809]: I1205 12:44:09.242614 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6974cc94df-frb66" event={"ID":"9da25051-70b9-4014-b10c-900022a35a1a","Type":"ContainerDied","Data":"0c311f82c9edb7b608b237a67d4bd4e751e5f1ce3fd4a92d8f4043b66d73021e"} Dec 05 12:44:09 crc kubenswrapper[4809]: I1205 12:44:09.242660 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6974cc94df-frb66" event={"ID":"9da25051-70b9-4014-b10c-900022a35a1a","Type":"ContainerStarted","Data":"1ed84707883bb40c4f407034fbec26fdd29d63025ba2b3164379aaf6a06f5729"} Dec 05 12:44:10 crc kubenswrapper[4809]: I1205 12:44:10.254921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"38ef0e35-a057-43e9-901d-e02be6abcca7","Type":"ContainerStarted","Data":"b9a73c767c2e152f69be560b1810464f231616fceb537886aa5b2f88de292254"} Dec 05 12:44:10 crc kubenswrapper[4809]: I1205 12:44:10.255277 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 12:44:10 crc kubenswrapper[4809]: I1205 12:44:10.257157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6974cc94df-frb66" event={"ID":"9da25051-70b9-4014-b10c-900022a35a1a","Type":"ContainerStarted","Data":"ddcbe5ce7b0f365a6b07d0da94b0da1a31dbf33569aebaa430d5cae3be051a4c"} Dec 05 12:44:10 crc kubenswrapper[4809]: I1205 12:44:10.257331 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:10 crc kubenswrapper[4809]: I1205 12:44:10.282483 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.282456147 podStartE2EDuration="3.282456147s" podCreationTimestamp="2025-12-05 12:44:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:10.270207468 +0000 UTC m=+5745.661184056" watchObservedRunningTime="2025-12-05 12:44:10.282456147 +0000 UTC m=+5745.673432705" Dec 05 12:44:10 crc kubenswrapper[4809]: I1205 12:44:10.294357 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6974cc94df-frb66" podStartSLOduration=3.294335427 podStartE2EDuration="3.294335427s" podCreationTimestamp="2025-12-05 12:44:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:10.290532755 +0000 UTC m=+5745.681509323" watchObservedRunningTime="2025-12-05 12:44:10.294335427 +0000 UTC m=+5745.685311985" Dec 05 12:44:14 crc kubenswrapper[4809]: E1205 12:44:14.943406 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f6967d_1d38_42a1_83de_5a3f220b5cb8.slice/crio-c4165bd85cdee484091f5d6dec2081b8ec3ef13e7387ad41a86cbbf81fbbedb8\": RecentStats: unable to find data in memory cache]" Dec 05 12:44:16 crc kubenswrapper[4809]: I1205 12:44:16.872223 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:44:16 crc kubenswrapper[4809]: E1205 12:44:16.872863 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:44:17 crc kubenswrapper[4809]: I1205 12:44:17.829843 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:44:17 crc kubenswrapper[4809]: I1205 12:44:17.918229 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-858bd85b85-6f4h2"] Dec 05 12:44:17 crc kubenswrapper[4809]: I1205 12:44:17.918455 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" podUID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerName="dnsmasq-dns" containerID="cri-o://77dd2145995dbf3d1876b48cc6df8ba035b66367f40ce343ee99670367794002" gracePeriod=10 Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.343464 4809 generic.go:334] "Generic (PLEG): container finished" podID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerID="77dd2145995dbf3d1876b48cc6df8ba035b66367f40ce343ee99670367794002" exitCode=0 Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.343992 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" event={"ID":"07125729-91e5-46ca-a996-a8c24cd60ab8","Type":"ContainerDied","Data":"77dd2145995dbf3d1876b48cc6df8ba035b66367f40ce343ee99670367794002"} Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.505165 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.635378 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-config\") pod \"07125729-91e5-46ca-a996-a8c24cd60ab8\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.635559 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-dns-svc\") pod \"07125729-91e5-46ca-a996-a8c24cd60ab8\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.635599 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-nb\") pod \"07125729-91e5-46ca-a996-a8c24cd60ab8\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.636384 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p46wt\" (UniqueName: \"kubernetes.io/projected/07125729-91e5-46ca-a996-a8c24cd60ab8-kube-api-access-p46wt\") pod \"07125729-91e5-46ca-a996-a8c24cd60ab8\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.636424 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-sb\") pod \"07125729-91e5-46ca-a996-a8c24cd60ab8\" (UID: \"07125729-91e5-46ca-a996-a8c24cd60ab8\") " Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.659727 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07125729-91e5-46ca-a996-a8c24cd60ab8-kube-api-access-p46wt" (OuterVolumeSpecName: "kube-api-access-p46wt") pod "07125729-91e5-46ca-a996-a8c24cd60ab8" (UID: "07125729-91e5-46ca-a996-a8c24cd60ab8"). InnerVolumeSpecName "kube-api-access-p46wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.691380 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "07125729-91e5-46ca-a996-a8c24cd60ab8" (UID: "07125729-91e5-46ca-a996-a8c24cd60ab8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.700148 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "07125729-91e5-46ca-a996-a8c24cd60ab8" (UID: "07125729-91e5-46ca-a996-a8c24cd60ab8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.700525 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-config" (OuterVolumeSpecName: "config") pod "07125729-91e5-46ca-a996-a8c24cd60ab8" (UID: "07125729-91e5-46ca-a996-a8c24cd60ab8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.708671 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "07125729-91e5-46ca-a996-a8c24cd60ab8" (UID: "07125729-91e5-46ca-a996-a8c24cd60ab8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.738196 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p46wt\" (UniqueName: \"kubernetes.io/projected/07125729-91e5-46ca-a996-a8c24cd60ab8-kube-api-access-p46wt\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.738229 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.738241 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.738249 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:18 crc kubenswrapper[4809]: I1205 12:44:18.738257 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07125729-91e5-46ca-a996-a8c24cd60ab8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.355210 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" event={"ID":"07125729-91e5-46ca-a996-a8c24cd60ab8","Type":"ContainerDied","Data":"146014f3a9ad620720fbabad77325a596414010cf90155c41977e48d0b3cff08"} Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.355250 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858bd85b85-6f4h2" Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.355549 4809 scope.go:117] "RemoveContainer" containerID="77dd2145995dbf3d1876b48cc6df8ba035b66367f40ce343ee99670367794002" Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.385356 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-858bd85b85-6f4h2"] Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.393878 4809 scope.go:117] "RemoveContainer" containerID="1f24312485989194c8523337b87143359da0d902a5d7d4601ce2b4e43cb92953" Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.410232 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-858bd85b85-6f4h2"] Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.486545 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.487944 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="74c9484e-89ae-4abf-9cdc-2953bd18bf8c" containerName="nova-cell0-conductor-conductor" containerID="cri-o://1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" gracePeriod=30 Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.531654 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.531915 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-log" containerID="cri-o://d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a" gracePeriod=30 Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.532324 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-metadata" containerID="cri-o://f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41" gracePeriod=30 Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.544766 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.545065 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="408b563d-0ca7-48c0-82a3-83683536d42a" containerName="nova-scheduler-scheduler" containerID="cri-o://0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5" gracePeriod=30 Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.553529 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.553779 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-log" containerID="cri-o://f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6" gracePeriod=30 Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.554337 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-api" containerID="cri-o://86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60" gracePeriod=30 Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.567386 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:44:19 crc kubenswrapper[4809]: I1205 12:44:19.567648 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bc68a01f-217d-4d14-bf85-05e3250c3a8e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d" gracePeriod=30 Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.124561 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.364852 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.366532 4809 generic.go:334] "Generic (PLEG): container finished" podID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerID="f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6" exitCode=143 Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.366620 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"080276e3-3e86-4d71-a8fc-c5d4927e7031","Type":"ContainerDied","Data":"f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6"} Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.368861 4809 generic.go:334] "Generic (PLEG): container finished" podID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerID="d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a" exitCode=143 Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.368918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f8dc59a-423c-4df6-b0a5-bf1842d1a269","Type":"ContainerDied","Data":"d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a"} Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.370149 4809 generic.go:334] "Generic (PLEG): container finished" podID="bc68a01f-217d-4d14-bf85-05e3250c3a8e" containerID="d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d" exitCode=0 Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.370174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bc68a01f-217d-4d14-bf85-05e3250c3a8e","Type":"ContainerDied","Data":"d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d"} Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.370188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bc68a01f-217d-4d14-bf85-05e3250c3a8e","Type":"ContainerDied","Data":"5bd258579ab0cf63dbc2d055f2960d5b5bfb05ed004f40d634e8ca70599ea62d"} Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.370204 4809 scope.go:117] "RemoveContainer" containerID="d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.370312 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.402436 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-combined-ca-bundle\") pod \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.402534 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-config-data\") pod \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.402709 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lxvk\" (UniqueName: \"kubernetes.io/projected/bc68a01f-217d-4d14-bf85-05e3250c3a8e-kube-api-access-8lxvk\") pod \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\" (UID: \"bc68a01f-217d-4d14-bf85-05e3250c3a8e\") " Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.414800 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc68a01f-217d-4d14-bf85-05e3250c3a8e-kube-api-access-8lxvk" (OuterVolumeSpecName: "kube-api-access-8lxvk") pod "bc68a01f-217d-4d14-bf85-05e3250c3a8e" (UID: "bc68a01f-217d-4d14-bf85-05e3250c3a8e"). InnerVolumeSpecName "kube-api-access-8lxvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.429829 4809 scope.go:117] "RemoveContainer" containerID="d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d" Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.430453 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.430799 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d\": container with ID starting with d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d not found: ID does not exist" containerID="d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.430908 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d"} err="failed to get container status \"d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d\": rpc error: code = NotFound desc = could not find container \"d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d\": container with ID starting with d07c830fbc2f4455acef878bb70c49d1bbb17114a044d11f241e1471b30e980d not found: ID does not exist" Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.433295 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.438893 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-config-data" (OuterVolumeSpecName: "config-data") pod "bc68a01f-217d-4d14-bf85-05e3250c3a8e" (UID: "bc68a01f-217d-4d14-bf85-05e3250c3a8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.439881 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.439957 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="408b563d-0ca7-48c0-82a3-83683536d42a" containerName="nova-scheduler-scheduler" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.457839 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc68a01f-217d-4d14-bf85-05e3250c3a8e" (UID: "bc68a01f-217d-4d14-bf85-05e3250c3a8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.504354 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.506094 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.506660 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.506707 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc68a01f-217d-4d14-bf85-05e3250c3a8e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.506722 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lxvk\" (UniqueName: \"kubernetes.io/projected/bc68a01f-217d-4d14-bf85-05e3250c3a8e-kube-api-access-8lxvk\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.509758 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.509868 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="74c9484e-89ae-4abf-9cdc-2953bd18bf8c" containerName="nova-cell0-conductor-conductor" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.705438 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.722404 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.742018 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.742743 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerName="dnsmasq-dns" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.742829 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerName="dnsmasq-dns" Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.742916 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerName="init" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.742980 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerName="init" Dec 05 12:44:20 crc kubenswrapper[4809]: E1205 12:44:20.743086 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc68a01f-217d-4d14-bf85-05e3250c3a8e" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.743167 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc68a01f-217d-4d14-bf85-05e3250c3a8e" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.743501 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="07125729-91e5-46ca-a996-a8c24cd60ab8" containerName="dnsmasq-dns" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.743630 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc68a01f-217d-4d14-bf85-05e3250c3a8e" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.744522 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.750088 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.750954 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.811088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0af04389-9c23-4925-aa2f-c0466ae00de3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.811221 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0af04389-9c23-4925-aa2f-c0466ae00de3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.811287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgv9l\" (UniqueName: \"kubernetes.io/projected/0af04389-9c23-4925-aa2f-c0466ae00de3-kube-api-access-lgv9l\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.888798 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07125729-91e5-46ca-a996-a8c24cd60ab8" path="/var/lib/kubelet/pods/07125729-91e5-46ca-a996-a8c24cd60ab8/volumes" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.889734 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc68a01f-217d-4d14-bf85-05e3250c3a8e" path="/var/lib/kubelet/pods/bc68a01f-217d-4d14-bf85-05e3250c3a8e/volumes" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.913401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0af04389-9c23-4925-aa2f-c0466ae00de3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.913469 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgv9l\" (UniqueName: \"kubernetes.io/projected/0af04389-9c23-4925-aa2f-c0466ae00de3-kube-api-access-lgv9l\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.913533 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0af04389-9c23-4925-aa2f-c0466ae00de3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.917232 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0af04389-9c23-4925-aa2f-c0466ae00de3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.919148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0af04389-9c23-4925-aa2f-c0466ae00de3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:20 crc kubenswrapper[4809]: I1205 12:44:20.929445 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgv9l\" (UniqueName: \"kubernetes.io/projected/0af04389-9c23-4925-aa2f-c0466ae00de3-kube-api-access-lgv9l\") pod \"nova-cell1-novncproxy-0\" (UID: \"0af04389-9c23-4925-aa2f-c0466ae00de3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:21 crc kubenswrapper[4809]: I1205 12:44:21.074795 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:21 crc kubenswrapper[4809]: I1205 12:44:21.541230 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 12:44:21 crc kubenswrapper[4809]: W1205 12:44:21.542231 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0af04389_9c23_4925_aa2f_c0466ae00de3.slice/crio-bc2878314390ee42159da960a78e012bfe9d8612cda68897cce6d1a082d8df93 WatchSource:0}: Error finding container bc2878314390ee42159da960a78e012bfe9d8612cda68897cce6d1a082d8df93: Status 404 returned error can't find the container with id bc2878314390ee42159da960a78e012bfe9d8612cda68897cce6d1a082d8df93 Dec 05 12:44:22 crc kubenswrapper[4809]: I1205 12:44:22.393724 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0af04389-9c23-4925-aa2f-c0466ae00de3","Type":"ContainerStarted","Data":"b0e5d18ebf3b0e2d7065b11d43aa6435e0ff90a82e5cdcebd0fa5173260901cf"} Dec 05 12:44:22 crc kubenswrapper[4809]: I1205 12:44:22.405490 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0af04389-9c23-4925-aa2f-c0466ae00de3","Type":"ContainerStarted","Data":"bc2878314390ee42159da960a78e012bfe9d8612cda68897cce6d1a082d8df93"} Dec 05 12:44:22 crc kubenswrapper[4809]: I1205 12:44:22.427398 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.427377576 podStartE2EDuration="2.427377576s" podCreationTimestamp="2025-12-05 12:44:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:22.425178207 +0000 UTC m=+5757.816154885" watchObservedRunningTime="2025-12-05 12:44:22.427377576 +0000 UTC m=+5757.818354134" Dec 05 12:44:22 crc kubenswrapper[4809]: I1205 12:44:22.758043 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:44:22 crc kubenswrapper[4809]: I1205 12:44:22.758545 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="c097deb0-a548-4b13-a41f-68fa3b98074a" containerName="nova-cell1-conductor-conductor" containerID="cri-o://cfac4fb786d41315a7e45dd20833f28176477f22c645dc48ff0c421df86b5681" gracePeriod=30 Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.153508 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.209757 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-combined-ca-bundle\") pod \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.209883 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-config-data\") pod \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.210024 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hckkn\" (UniqueName: \"kubernetes.io/projected/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-kube-api-access-hckkn\") pod \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.210088 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-logs\") pod \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\" (UID: \"2f8dc59a-423c-4df6-b0a5-bf1842d1a269\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.211174 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-logs" (OuterVolumeSpecName: "logs") pod "2f8dc59a-423c-4df6-b0a5-bf1842d1a269" (UID: "2f8dc59a-423c-4df6-b0a5-bf1842d1a269"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.216933 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-kube-api-access-hckkn" (OuterVolumeSpecName: "kube-api-access-hckkn") pod "2f8dc59a-423c-4df6-b0a5-bf1842d1a269" (UID: "2f8dc59a-423c-4df6-b0a5-bf1842d1a269"). InnerVolumeSpecName "kube-api-access-hckkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.220161 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.238847 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f8dc59a-423c-4df6-b0a5-bf1842d1a269" (UID: "2f8dc59a-423c-4df6-b0a5-bf1842d1a269"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.251968 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-config-data" (OuterVolumeSpecName: "config-data") pod "2f8dc59a-423c-4df6-b0a5-bf1842d1a269" (UID: "2f8dc59a-423c-4df6-b0a5-bf1842d1a269"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.311818 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-combined-ca-bundle\") pod \"080276e3-3e86-4d71-a8fc-c5d4927e7031\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.311913 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w9fg\" (UniqueName: \"kubernetes.io/projected/080276e3-3e86-4d71-a8fc-c5d4927e7031-kube-api-access-5w9fg\") pod \"080276e3-3e86-4d71-a8fc-c5d4927e7031\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.312042 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/080276e3-3e86-4d71-a8fc-c5d4927e7031-logs\") pod \"080276e3-3e86-4d71-a8fc-c5d4927e7031\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.312109 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-config-data\") pod \"080276e3-3e86-4d71-a8fc-c5d4927e7031\" (UID: \"080276e3-3e86-4d71-a8fc-c5d4927e7031\") " Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.312482 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.312501 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hckkn\" (UniqueName: \"kubernetes.io/projected/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-kube-api-access-hckkn\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.312513 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.312522 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8dc59a-423c-4df6-b0a5-bf1842d1a269-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.312652 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/080276e3-3e86-4d71-a8fc-c5d4927e7031-logs" (OuterVolumeSpecName: "logs") pod "080276e3-3e86-4d71-a8fc-c5d4927e7031" (UID: "080276e3-3e86-4d71-a8fc-c5d4927e7031"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.319123 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/080276e3-3e86-4d71-a8fc-c5d4927e7031-kube-api-access-5w9fg" (OuterVolumeSpecName: "kube-api-access-5w9fg") pod "080276e3-3e86-4d71-a8fc-c5d4927e7031" (UID: "080276e3-3e86-4d71-a8fc-c5d4927e7031"). InnerVolumeSpecName "kube-api-access-5w9fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.346941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-config-data" (OuterVolumeSpecName: "config-data") pod "080276e3-3e86-4d71-a8fc-c5d4927e7031" (UID: "080276e3-3e86-4d71-a8fc-c5d4927e7031"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.348956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "080276e3-3e86-4d71-a8fc-c5d4927e7031" (UID: "080276e3-3e86-4d71-a8fc-c5d4927e7031"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.404377 4809 generic.go:334] "Generic (PLEG): container finished" podID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerID="f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41" exitCode=0 Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.404430 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f8dc59a-423c-4df6-b0a5-bf1842d1a269","Type":"ContainerDied","Data":"f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41"} Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.404457 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f8dc59a-423c-4df6-b0a5-bf1842d1a269","Type":"ContainerDied","Data":"d5f8c0a77ddc747a14371da95677be32a728a90df2d633bc13eccc3c8e8474d4"} Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.404472 4809 scope.go:117] "RemoveContainer" containerID="f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.404586 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.410119 4809 generic.go:334] "Generic (PLEG): container finished" podID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerID="86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60" exitCode=0 Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.410568 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.410934 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"080276e3-3e86-4d71-a8fc-c5d4927e7031","Type":"ContainerDied","Data":"86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60"} Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.410967 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"080276e3-3e86-4d71-a8fc-c5d4927e7031","Type":"ContainerDied","Data":"b245d98bafd84e724512feea83dff46fb358914281b3c1d5055d17d2effb0bdd"} Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.414237 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/080276e3-3e86-4d71-a8fc-c5d4927e7031-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.414272 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.414286 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/080276e3-3e86-4d71-a8fc-c5d4927e7031-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.414301 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w9fg\" (UniqueName: \"kubernetes.io/projected/080276e3-3e86-4d71-a8fc-c5d4927e7031-kube-api-access-5w9fg\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.442733 4809 scope.go:117] "RemoveContainer" containerID="d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.480029 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.489455 4809 scope.go:117] "RemoveContainer" containerID="f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41" Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.492799 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41\": container with ID starting with f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41 not found: ID does not exist" containerID="f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.492856 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41"} err="failed to get container status \"f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41\": rpc error: code = NotFound desc = could not find container \"f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41\": container with ID starting with f6303a6fd925e0915d55d0fb1d7cc9ed3db3a2a47b0b598693d8fc21a3793f41 not found: ID does not exist" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.492889 4809 scope.go:117] "RemoveContainer" containerID="d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a" Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.495831 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a\": container with ID starting with d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a not found: ID does not exist" containerID="d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.495880 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a"} err="failed to get container status \"d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a\": rpc error: code = NotFound desc = could not find container \"d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a\": container with ID starting with d39e5a0945de6c2df8f4f79f08ca9794ecb6ec0e38cb1fddd0c39671caa78c9a not found: ID does not exist" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.495912 4809 scope.go:117] "RemoveContainer" containerID="86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.497761 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.506528 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.506977 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-log" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.506998 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-log" Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.507013 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-log" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.507018 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-log" Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.507035 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-metadata" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.507040 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-metadata" Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.507057 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-api" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.507064 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-api" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.507241 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-api" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.507257 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-metadata" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.507266 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-log" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.507280 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-log" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.508238 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.511857 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.519226 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.531338 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.543360 4809 scope.go:117] "RemoveContainer" containerID="f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.547043 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.549615 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.554441 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.556716 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.564041 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.584901 4809 scope.go:117] "RemoveContainer" containerID="86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60" Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.585603 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60\": container with ID starting with 86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60 not found: ID does not exist" containerID="86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.585654 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60"} err="failed to get container status \"86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60\": rpc error: code = NotFound desc = could not find container \"86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60\": container with ID starting with 86959b79f119f4ecb1cbd186d70b4ecea7ef031c47ae4195544701587e084f60 not found: ID does not exist" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.585682 4809 scope.go:117] "RemoveContainer" containerID="f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6" Dec 05 12:44:23 crc kubenswrapper[4809]: E1205 12:44:23.585962 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6\": container with ID starting with f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6 not found: ID does not exist" containerID="f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.585994 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6"} err="failed to get container status \"f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6\": rpc error: code = NotFound desc = could not find container \"f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6\": container with ID starting with f4a747e54939a17be7353db9cc98690cd400a039041448125b486681deec65d6 not found: ID does not exist" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631746 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-config-data\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631814 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d5dz\" (UniqueName: \"kubernetes.io/projected/b405ddbd-2a5f-4126-8c54-4bb55754c432-kube-api-access-2d5dz\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631850 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf8cv\" (UniqueName: \"kubernetes.io/projected/51d9c44b-44e8-4bd6-876c-2047efc32605-kube-api-access-rf8cv\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631875 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631900 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-config-data\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631926 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631948 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51d9c44b-44e8-4bd6-876c-2047efc32605-logs\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.631976 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b405ddbd-2a5f-4126-8c54-4bb55754c432-logs\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733590 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733669 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-config-data\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733699 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733724 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51d9c44b-44e8-4bd6-876c-2047efc32605-logs\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733752 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b405ddbd-2a5f-4126-8c54-4bb55754c432-logs\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733812 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-config-data\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d5dz\" (UniqueName: \"kubernetes.io/projected/b405ddbd-2a5f-4126-8c54-4bb55754c432-kube-api-access-2d5dz\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.733887 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf8cv\" (UniqueName: \"kubernetes.io/projected/51d9c44b-44e8-4bd6-876c-2047efc32605-kube-api-access-rf8cv\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.735380 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51d9c44b-44e8-4bd6-876c-2047efc32605-logs\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.735534 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b405ddbd-2a5f-4126-8c54-4bb55754c432-logs\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.737883 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.738077 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.738766 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-config-data\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.745446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-config-data\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.754893 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d5dz\" (UniqueName: \"kubernetes.io/projected/b405ddbd-2a5f-4126-8c54-4bb55754c432-kube-api-access-2d5dz\") pod \"nova-api-0\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " pod="openstack/nova-api-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.759265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf8cv\" (UniqueName: \"kubernetes.io/projected/51d9c44b-44e8-4bd6-876c-2047efc32605-kube-api-access-rf8cv\") pod \"nova-metadata-0\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.838116 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 12:44:23 crc kubenswrapper[4809]: I1205 12:44:23.875211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.429524 4809 generic.go:334] "Generic (PLEG): container finished" podID="c097deb0-a548-4b13-a41f-68fa3b98074a" containerID="cfac4fb786d41315a7e45dd20833f28176477f22c645dc48ff0c421df86b5681" exitCode=0 Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.430155 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c097deb0-a548-4b13-a41f-68fa3b98074a","Type":"ContainerDied","Data":"cfac4fb786d41315a7e45dd20833f28176477f22c645dc48ff0c421df86b5681"} Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.433426 4809 generic.go:334] "Generic (PLEG): container finished" podID="408b563d-0ca7-48c0-82a3-83683536d42a" containerID="0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5" exitCode=0 Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.433477 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"408b563d-0ca7-48c0-82a3-83683536d42a","Type":"ContainerDied","Data":"0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5"} Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.505489 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.586437 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.665250 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.668330 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-combined-ca-bundle\") pod \"408b563d-0ca7-48c0-82a3-83683536d42a\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.668496 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-config-data\") pod \"408b563d-0ca7-48c0-82a3-83683536d42a\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.668600 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdgs4\" (UniqueName: \"kubernetes.io/projected/408b563d-0ca7-48c0-82a3-83683536d42a-kube-api-access-kdgs4\") pod \"408b563d-0ca7-48c0-82a3-83683536d42a\" (UID: \"408b563d-0ca7-48c0-82a3-83683536d42a\") " Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.679553 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/408b563d-0ca7-48c0-82a3-83683536d42a-kube-api-access-kdgs4" (OuterVolumeSpecName: "kube-api-access-kdgs4") pod "408b563d-0ca7-48c0-82a3-83683536d42a" (UID: "408b563d-0ca7-48c0-82a3-83683536d42a"). InnerVolumeSpecName "kube-api-access-kdgs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.700732 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-config-data" (OuterVolumeSpecName: "config-data") pod "408b563d-0ca7-48c0-82a3-83683536d42a" (UID: "408b563d-0ca7-48c0-82a3-83683536d42a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.712754 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.721073 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "408b563d-0ca7-48c0-82a3-83683536d42a" (UID: "408b563d-0ca7-48c0-82a3-83683536d42a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.772504 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.772529 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdgs4\" (UniqueName: \"kubernetes.io/projected/408b563d-0ca7-48c0-82a3-83683536d42a-kube-api-access-kdgs4\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.772541 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408b563d-0ca7-48c0-82a3-83683536d42a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.879591 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-config-data\") pod \"c097deb0-a548-4b13-a41f-68fa3b98074a\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.879684 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5d8p\" (UniqueName: \"kubernetes.io/projected/c097deb0-a548-4b13-a41f-68fa3b98074a-kube-api-access-j5d8p\") pod \"c097deb0-a548-4b13-a41f-68fa3b98074a\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.879755 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-combined-ca-bundle\") pod \"c097deb0-a548-4b13-a41f-68fa3b98074a\" (UID: \"c097deb0-a548-4b13-a41f-68fa3b98074a\") " Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.888810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c097deb0-a548-4b13-a41f-68fa3b98074a-kube-api-access-j5d8p" (OuterVolumeSpecName: "kube-api-access-j5d8p") pod "c097deb0-a548-4b13-a41f-68fa3b98074a" (UID: "c097deb0-a548-4b13-a41f-68fa3b98074a"). InnerVolumeSpecName "kube-api-access-j5d8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.891740 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5d8p\" (UniqueName: \"kubernetes.io/projected/c097deb0-a548-4b13-a41f-68fa3b98074a-kube-api-access-j5d8p\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.899907 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" path="/var/lib/kubelet/pods/080276e3-3e86-4d71-a8fc-c5d4927e7031/volumes" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.900871 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" path="/var/lib/kubelet/pods/2f8dc59a-423c-4df6-b0a5-bf1842d1a269/volumes" Dec 05 12:44:24 crc kubenswrapper[4809]: E1205 12:44:24.903856 4809 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/3e64435bffcd9e23050c29276bb23ff1a7bde810fab5e37f05c6d78dc5ad8e4f/diff" to get inode usage: stat /var/lib/containers/storage/overlay/3e64435bffcd9e23050c29276bb23ff1a7bde810fab5e37f05c6d78dc5ad8e4f/diff: no such file or directory, extraDiskErr: Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.918160 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c097deb0-a548-4b13-a41f-68fa3b98074a" (UID: "c097deb0-a548-4b13-a41f-68fa3b98074a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.948244 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-config-data" (OuterVolumeSpecName: "config-data") pod "c097deb0-a548-4b13-a41f-68fa3b98074a" (UID: "c097deb0-a548-4b13-a41f-68fa3b98074a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.992804 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:24 crc kubenswrapper[4809]: I1205 12:44:24.993131 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c097deb0-a548-4b13-a41f-68fa3b98074a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.049716 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.109482 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-combined-ca-bundle\") pod \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.109595 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqfnr\" (UniqueName: \"kubernetes.io/projected/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-kube-api-access-sqfnr\") pod \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.109701 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-config-data\") pod \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\" (UID: \"74c9484e-89ae-4abf-9cdc-2953bd18bf8c\") " Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.120392 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-kube-api-access-sqfnr" (OuterVolumeSpecName: "kube-api-access-sqfnr") pod "74c9484e-89ae-4abf-9cdc-2953bd18bf8c" (UID: "74c9484e-89ae-4abf-9cdc-2953bd18bf8c"). InnerVolumeSpecName "kube-api-access-sqfnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.157093 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74c9484e-89ae-4abf-9cdc-2953bd18bf8c" (UID: "74c9484e-89ae-4abf-9cdc-2953bd18bf8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.157180 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-config-data" (OuterVolumeSpecName: "config-data") pod "74c9484e-89ae-4abf-9cdc-2953bd18bf8c" (UID: "74c9484e-89ae-4abf-9cdc-2953bd18bf8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.211357 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.211420 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqfnr\" (UniqueName: \"kubernetes.io/projected/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-kube-api-access-sqfnr\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.211433 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9484e-89ae-4abf-9cdc-2953bd18bf8c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:25 crc kubenswrapper[4809]: E1205 12:44:25.219930 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod408b563d_0ca7_48c0_82a3_83683536d42a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod408b563d_0ca7_48c0_82a3_83683536d42a.slice/crio-e63a36c58725cc0e682b94283bf513aba2a8d6a004c3c2551c2e2e0b9ee34343\": RecentStats: unable to find data in memory cache]" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.450570 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c097deb0-a548-4b13-a41f-68fa3b98074a","Type":"ContainerDied","Data":"dd51f625902330dba56cde9760d25bed0bd328988c68e5f778162edd8176f1d1"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.450727 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.450918 4809 scope.go:117] "RemoveContainer" containerID="cfac4fb786d41315a7e45dd20833f28176477f22c645dc48ff0c421df86b5681" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.453278 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"408b563d-0ca7-48c0-82a3-83683536d42a","Type":"ContainerDied","Data":"e63a36c58725cc0e682b94283bf513aba2a8d6a004c3c2551c2e2e0b9ee34343"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.453751 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.463827 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b405ddbd-2a5f-4126-8c54-4bb55754c432","Type":"ContainerStarted","Data":"4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.463881 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b405ddbd-2a5f-4126-8c54-4bb55754c432","Type":"ContainerStarted","Data":"4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.463896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b405ddbd-2a5f-4126-8c54-4bb55754c432","Type":"ContainerStarted","Data":"628e49c1023e0e868b0f8979616333eee6251bc8cc72f73847542f0a40d5d798"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.465783 4809 generic.go:334] "Generic (PLEG): container finished" podID="74c9484e-89ae-4abf-9cdc-2953bd18bf8c" containerID="1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" exitCode=0 Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.465855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"74c9484e-89ae-4abf-9cdc-2953bd18bf8c","Type":"ContainerDied","Data":"1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.465876 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"74c9484e-89ae-4abf-9cdc-2953bd18bf8c","Type":"ContainerDied","Data":"03efaa4cf4a06b3d49dd9e39fe1c07c13e31919c10370d3409a7360ab9e2c631"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.465889 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.467347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51d9c44b-44e8-4bd6-876c-2047efc32605","Type":"ContainerStarted","Data":"685061eb6291332a3626770b6c95c85cbdd8d9c9d41ad36969519a969ec3e998"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.467394 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51d9c44b-44e8-4bd6-876c-2047efc32605","Type":"ContainerStarted","Data":"7330f21737c9df36e111348a792051200c138724d80a08c38a1d8dcbab55b678"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.467408 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51d9c44b-44e8-4bd6-876c-2047efc32605","Type":"ContainerStarted","Data":"4b756ba9cea93acd2ba9b7d66bd8e2ef9597fab25831b824ba004cb8f6f82ea3"} Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.480843 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.486439 4809 scope.go:117] "RemoveContainer" containerID="0833c13428b983fba3c1fda8920f10dc1d687ea528b2bd088fbafc541e503dd5" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.502596 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.522090 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: E1205 12:44:25.522551 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c9484e-89ae-4abf-9cdc-2953bd18bf8c" containerName="nova-cell0-conductor-conductor" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.522572 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c9484e-89ae-4abf-9cdc-2953bd18bf8c" containerName="nova-cell0-conductor-conductor" Dec 05 12:44:25 crc kubenswrapper[4809]: E1205 12:44:25.522594 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408b563d-0ca7-48c0-82a3-83683536d42a" containerName="nova-scheduler-scheduler" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.522602 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="408b563d-0ca7-48c0-82a3-83683536d42a" containerName="nova-scheduler-scheduler" Dec 05 12:44:25 crc kubenswrapper[4809]: E1205 12:44:25.522616 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c097deb0-a548-4b13-a41f-68fa3b98074a" containerName="nova-cell1-conductor-conductor" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.522624 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c097deb0-a548-4b13-a41f-68fa3b98074a" containerName="nova-cell1-conductor-conductor" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.522856 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c097deb0-a548-4b13-a41f-68fa3b98074a" containerName="nova-cell1-conductor-conductor" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.522884 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c9484e-89ae-4abf-9cdc-2953bd18bf8c" containerName="nova-cell0-conductor-conductor" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.522911 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="408b563d-0ca7-48c0-82a3-83683536d42a" containerName="nova-scheduler-scheduler" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.523582 4809 scope.go:117] "RemoveContainer" containerID="1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.523734 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.535086 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.562194 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.562167204 podStartE2EDuration="2.562167204s" podCreationTimestamp="2025-12-05 12:44:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:25.504877751 +0000 UTC m=+5760.895854329" watchObservedRunningTime="2025-12-05 12:44:25.562167204 +0000 UTC m=+5760.953143772" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.589606 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.593756 4809 scope.go:117] "RemoveContainer" containerID="1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" Dec 05 12:44:25 crc kubenswrapper[4809]: E1205 12:44:25.594198 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23\": container with ID starting with 1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23 not found: ID does not exist" containerID="1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.594228 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23"} err="failed to get container status \"1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23\": rpc error: code = NotFound desc = could not find container \"1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23\": container with ID starting with 1d4a435359848452a98783bdcdf618b95d9ffbb4dd51c2b968cc252ff0f77d23 not found: ID does not exist" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.595715 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.595691156 podStartE2EDuration="2.595691156s" podCreationTimestamp="2025-12-05 12:44:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:25.543545152 +0000 UTC m=+5760.934521710" watchObservedRunningTime="2025-12-05 12:44:25.595691156 +0000 UTC m=+5760.986667714" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.607688 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.626720 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.637299 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.638826 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.642024 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.676857 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.685733 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.693491 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.701787 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.703162 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.705361 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.710731 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.718980 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.719150 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-config-data\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.719221 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqtff\" (UniqueName: \"kubernetes.io/projected/49c8cdbe-f1e8-4d73-a86f-d18399642b85-kube-api-access-bqtff\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821562 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwgjn\" (UniqueName: \"kubernetes.io/projected/daf4bee5-c20a-4a5a-b6f7-df566b286601-kube-api-access-qwgjn\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821605 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821664 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821704 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821742 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgvvw\" (UniqueName: \"kubernetes.io/projected/6644e4f7-66da-4ab0-8100-18ba3cac40bf-kube-api-access-jgvvw\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821772 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-config-data\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqtff\" (UniqueName: \"kubernetes.io/projected/49c8cdbe-f1e8-4d73-a86f-d18399642b85-kube-api-access-bqtff\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821821 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.821849 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.827541 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-config-data\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.831000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.841197 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqtff\" (UniqueName: \"kubernetes.io/projected/49c8cdbe-f1e8-4d73-a86f-d18399642b85-kube-api-access-bqtff\") pod \"nova-scheduler-0\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.864287 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.924132 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwgjn\" (UniqueName: \"kubernetes.io/projected/daf4bee5-c20a-4a5a-b6f7-df566b286601-kube-api-access-qwgjn\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.924498 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.924538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.924593 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgvvw\" (UniqueName: \"kubernetes.io/projected/6644e4f7-66da-4ab0-8100-18ba3cac40bf-kube-api-access-jgvvw\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.924661 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.924684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.929075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.929245 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.929283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.930328 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.952918 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwgjn\" (UniqueName: \"kubernetes.io/projected/daf4bee5-c20a-4a5a-b6f7-df566b286601-kube-api-access-qwgjn\") pod \"nova-cell0-conductor-0\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.958254 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgvvw\" (UniqueName: \"kubernetes.io/projected/6644e4f7-66da-4ab0-8100-18ba3cac40bf-kube-api-access-jgvvw\") pod \"nova-cell1-conductor-0\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:25 crc kubenswrapper[4809]: I1205 12:44:25.967411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.018317 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.075322 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.356867 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 12:44:26 crc kubenswrapper[4809]: W1205 12:44:26.363678 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49c8cdbe_f1e8_4d73_a86f_d18399642b85.slice/crio-84155bcfb21d37c0ad797cc4a9373c633d4dc8a0a903408018d210af5481143f WatchSource:0}: Error finding container 84155bcfb21d37c0ad797cc4a9373c633d4dc8a0a903408018d210af5481143f: Status 404 returned error can't find the container with id 84155bcfb21d37c0ad797cc4a9373c633d4dc8a0a903408018d210af5481143f Dec 05 12:44:26 crc kubenswrapper[4809]: W1205 12:44:26.448168 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6644e4f7_66da_4ab0_8100_18ba3cac40bf.slice/crio-3c34cd74a0a8fc04620d0abc4d63e8a74632cc2cacfc5316f9a2fa8e1d0c28f5 WatchSource:0}: Error finding container 3c34cd74a0a8fc04620d0abc4d63e8a74632cc2cacfc5316f9a2fa8e1d0c28f5: Status 404 returned error can't find the container with id 3c34cd74a0a8fc04620d0abc4d63e8a74632cc2cacfc5316f9a2fa8e1d0c28f5 Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.448506 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.491916 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6644e4f7-66da-4ab0-8100-18ba3cac40bf","Type":"ContainerStarted","Data":"3c34cd74a0a8fc04620d0abc4d63e8a74632cc2cacfc5316f9a2fa8e1d0c28f5"} Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.492916 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49c8cdbe-f1e8-4d73-a86f-d18399642b85","Type":"ContainerStarted","Data":"84155bcfb21d37c0ad797cc4a9373c633d4dc8a0a903408018d210af5481143f"} Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.540831 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 12:44:26 crc kubenswrapper[4809]: W1205 12:44:26.543509 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf4bee5_c20a_4a5a_b6f7_df566b286601.slice/crio-a1e6c813fa8c2a540a5bd9ef067cf6bb8db7dd5f9d11893185a68459b76b0fed WatchSource:0}: Error finding container a1e6c813fa8c2a540a5bd9ef067cf6bb8db7dd5f9d11893185a68459b76b0fed: Status 404 returned error can't find the container with id a1e6c813fa8c2a540a5bd9ef067cf6bb8db7dd5f9d11893185a68459b76b0fed Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.883889 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="408b563d-0ca7-48c0-82a3-83683536d42a" path="/var/lib/kubelet/pods/408b563d-0ca7-48c0-82a3-83683536d42a/volumes" Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.884835 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74c9484e-89ae-4abf-9cdc-2953bd18bf8c" path="/var/lib/kubelet/pods/74c9484e-89ae-4abf-9cdc-2953bd18bf8c/volumes" Dec 05 12:44:26 crc kubenswrapper[4809]: I1205 12:44:26.885304 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c097deb0-a548-4b13-a41f-68fa3b98074a" path="/var/lib/kubelet/pods/c097deb0-a548-4b13-a41f-68fa3b98074a/volumes" Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.502602 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"daf4bee5-c20a-4a5a-b6f7-df566b286601","Type":"ContainerStarted","Data":"9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec"} Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.502676 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"daf4bee5-c20a-4a5a-b6f7-df566b286601","Type":"ContainerStarted","Data":"a1e6c813fa8c2a540a5bd9ef067cf6bb8db7dd5f9d11893185a68459b76b0fed"} Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.503030 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.515834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6644e4f7-66da-4ab0-8100-18ba3cac40bf","Type":"ContainerStarted","Data":"1ad11e73e4564fb1f0f0533d92ad600562ff60c4270a24cf1f1115b427f4097a"} Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.516369 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.524679 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49c8cdbe-f1e8-4d73-a86f-d18399642b85","Type":"ContainerStarted","Data":"db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658"} Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.531648 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.53160533 podStartE2EDuration="2.53160533s" podCreationTimestamp="2025-12-05 12:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:27.522270049 +0000 UTC m=+5762.913246617" watchObservedRunningTime="2025-12-05 12:44:27.53160533 +0000 UTC m=+5762.922581898" Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.565955 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.565931634 podStartE2EDuration="2.565931634s" podCreationTimestamp="2025-12-05 12:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:27.542681738 +0000 UTC m=+5762.933658306" watchObservedRunningTime="2025-12-05 12:44:27.565931634 +0000 UTC m=+5762.956908202" Dec 05 12:44:27 crc kubenswrapper[4809]: I1205 12:44:27.568977 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.568969306 podStartE2EDuration="2.568969306s" podCreationTimestamp="2025-12-05 12:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:27.563988482 +0000 UTC m=+5762.954965040" watchObservedRunningTime="2025-12-05 12:44:27.568969306 +0000 UTC m=+5762.959945874" Dec 05 12:44:28 crc kubenswrapper[4809]: I1205 12:44:28.109119 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:28 crc kubenswrapper[4809]: I1205 12:44:28.109560 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2f8dc59a-423c-4df6-b0a5-bf1842d1a269" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:28 crc kubenswrapper[4809]: I1205 12:44:28.839618 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:44:28 crc kubenswrapper[4809]: I1205 12:44:28.839692 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 12:44:30 crc kubenswrapper[4809]: I1205 12:44:30.865790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 12:44:31 crc kubenswrapper[4809]: I1205 12:44:31.045622 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 12:44:31 crc kubenswrapper[4809]: I1205 12:44:31.075338 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:31 crc kubenswrapper[4809]: I1205 12:44:31.092723 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:31 crc kubenswrapper[4809]: I1205 12:44:31.570694 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 12:44:31 crc kubenswrapper[4809]: I1205 12:44:31.872352 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:44:31 crc kubenswrapper[4809]: E1205 12:44:31.872739 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:44:33 crc kubenswrapper[4809]: I1205 12:44:33.839106 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 12:44:33 crc kubenswrapper[4809]: I1205 12:44:33.840361 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 12:44:33 crc kubenswrapper[4809]: I1205 12:44:33.876255 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 12:44:33 crc kubenswrapper[4809]: I1205 12:44:33.876313 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 12:44:35 crc kubenswrapper[4809]: I1205 12:44:35.003797 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.83:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:35 crc kubenswrapper[4809]: I1205 12:44:35.003828 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.84:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:35 crc kubenswrapper[4809]: I1205 12:44:35.003828 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.83:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:35 crc kubenswrapper[4809]: I1205 12:44:35.003889 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.84:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:35 crc kubenswrapper[4809]: I1205 12:44:35.865064 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 12:44:35 crc kubenswrapper[4809]: I1205 12:44:35.906277 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 12:44:36 crc kubenswrapper[4809]: I1205 12:44:36.000543 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 12:44:36 crc kubenswrapper[4809]: I1205 12:44:36.647185 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.219469 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.222118 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.224890 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.247576 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.357009 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.357046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg48w\" (UniqueName: \"kubernetes.io/projected/e11bde65-d1a0-46a0-897a-573299b21ad1-kube-api-access-cg48w\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.357246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-scripts\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.357528 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.357772 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e11bde65-d1a0-46a0-897a-573299b21ad1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.357829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.459189 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.459223 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg48w\" (UniqueName: \"kubernetes.io/projected/e11bde65-d1a0-46a0-897a-573299b21ad1-kube-api-access-cg48w\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.459265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-scripts\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.459330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.459384 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e11bde65-d1a0-46a0-897a-573299b21ad1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.459407 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.459585 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e11bde65-d1a0-46a0-897a-573299b21ad1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.465353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-scripts\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.465505 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.465544 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.467047 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.476238 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg48w\" (UniqueName: \"kubernetes.io/projected/e11bde65-d1a0-46a0-897a-573299b21ad1-kube-api-access-cg48w\") pod \"cinder-scheduler-0\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:38 crc kubenswrapper[4809]: I1205 12:44:38.551947 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 12:44:39 crc kubenswrapper[4809]: I1205 12:44:38.999072 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:39 crc kubenswrapper[4809]: W1205 12:44:38.999501 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode11bde65_d1a0_46a0_897a_573299b21ad1.slice/crio-28473b30787b940b8e3ffd5fa29923f45c30b40e2b5e44e67f47bd977bdc7d13 WatchSource:0}: Error finding container 28473b30787b940b8e3ffd5fa29923f45c30b40e2b5e44e67f47bd977bdc7d13: Status 404 returned error can't find the container with id 28473b30787b940b8e3ffd5fa29923f45c30b40e2b5e44e67f47bd977bdc7d13 Dec 05 12:44:39 crc kubenswrapper[4809]: I1205 12:44:39.634258 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:39 crc kubenswrapper[4809]: I1205 12:44:39.634766 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api-log" containerID="cri-o://fbd933c23104016a5dbdbd261a9c6c40593e4cf9ec15ee5cdeaed66e05ae6a7e" gracePeriod=30 Dec 05 12:44:39 crc kubenswrapper[4809]: I1205 12:44:39.634874 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api" containerID="cri-o://b9a73c767c2e152f69be560b1810464f231616fceb537886aa5b2f88de292254" gracePeriod=30 Dec 05 12:44:39 crc kubenswrapper[4809]: I1205 12:44:39.657357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e11bde65-d1a0-46a0-897a-573299b21ad1","Type":"ContainerStarted","Data":"0063b9770b72d2c09db64a01e4b016123ebea9d1d0b2902a96a0a48be6eee09c"} Dec 05 12:44:39 crc kubenswrapper[4809]: I1205 12:44:39.657411 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e11bde65-d1a0-46a0-897a-573299b21ad1","Type":"ContainerStarted","Data":"28473b30787b940b8e3ffd5fa29923f45c30b40e2b5e44e67f47bd977bdc7d13"} Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.388159 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.390357 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.399519 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.403072 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505652 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505762 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505795 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505818 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505847 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blbjt\" (UniqueName: \"kubernetes.io/projected/2f939aa0-d8aa-4771-a4fe-326db3a407df-kube-api-access-blbjt\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505899 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-run\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505932 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2f939aa0-d8aa-4771-a4fe-326db3a407df-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.505997 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.506023 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.506049 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.506072 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.506102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.506132 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.506173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607374 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607435 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607465 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blbjt\" (UniqueName: \"kubernetes.io/projected/2f939aa0-d8aa-4771-a4fe-326db3a407df-kube-api-access-blbjt\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607509 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-run\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2f939aa0-d8aa-4771-a4fe-326db3a407df-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607585 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607645 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607662 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607658 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-run\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607731 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607763 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607830 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607831 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607888 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607915 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607987 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.608005 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.607995 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2f939aa0-d8aa-4771-a4fe-326db3a407df-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.613526 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.614211 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.615437 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.618668 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2f939aa0-d8aa-4771-a4fe-326db3a407df-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.625615 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f939aa0-d8aa-4771-a4fe-326db3a407df-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.641542 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blbjt\" (UniqueName: \"kubernetes.io/projected/2f939aa0-d8aa-4771-a4fe-326db3a407df-kube-api-access-blbjt\") pod \"cinder-volume-volume1-0\" (UID: \"2f939aa0-d8aa-4771-a4fe-326db3a407df\") " pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.670801 4809 generic.go:334] "Generic (PLEG): container finished" podID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerID="fbd933c23104016a5dbdbd261a9c6c40593e4cf9ec15ee5cdeaed66e05ae6a7e" exitCode=143 Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.670878 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"38ef0e35-a057-43e9-901d-e02be6abcca7","Type":"ContainerDied","Data":"fbd933c23104016a5dbdbd261a9c6c40593e4cf9ec15ee5cdeaed66e05ae6a7e"} Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.673185 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e11bde65-d1a0-46a0-897a-573299b21ad1","Type":"ContainerStarted","Data":"d5f26fb137cfe1d96eb5ba1df0c71f711eff14633bc199cad440c9cb3fd07d39"} Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.702438 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.702419296 podStartE2EDuration="2.702419296s" podCreationTimestamp="2025-12-05 12:44:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:40.691367969 +0000 UTC m=+5776.082344537" watchObservedRunningTime="2025-12-05 12:44:40.702419296 +0000 UTC m=+5776.093395844" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.712791 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.950904 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.958739 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.958855 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 05 12:44:40 crc kubenswrapper[4809]: I1205 12:44:40.963598 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.122901 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-scripts\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.122952 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e41e2028-117d-4e52-878b-0c722a1891aa-ceph\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.122973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123007 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123040 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123060 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-config-data\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123079 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-sys\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123097 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123127 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-dev\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123164 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123177 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-run\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123192 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123235 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123267 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-lib-modules\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.123283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th7n2\" (UniqueName: \"kubernetes.io/projected/e41e2028-117d-4e52-878b-0c722a1891aa-kube-api-access-th7n2\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-lib-modules\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th7n2\" (UniqueName: \"kubernetes.io/projected/e41e2028-117d-4e52-878b-0c722a1891aa-kube-api-access-th7n2\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225126 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-scripts\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225148 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e41e2028-117d-4e52-878b-0c722a1891aa-ceph\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225239 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225270 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-config-data\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-sys\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225347 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-dev\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225392 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-run\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225675 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.225722 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-lib-modules\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.226882 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-sys\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.231178 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e41e2028-117d-4e52-878b-0c722a1891aa-ceph\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.231419 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.231457 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.231484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.234066 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-scripts\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.235069 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-config-data\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.235115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.235138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-dev\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.236996 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.237037 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-run\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.237063 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e41e2028-117d-4e52-878b-0c722a1891aa-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.238234 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e41e2028-117d-4e52-878b-0c722a1891aa-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.267318 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th7n2\" (UniqueName: \"kubernetes.io/projected/e41e2028-117d-4e52-878b-0c722a1891aa-kube-api-access-th7n2\") pod \"cinder-backup-0\" (UID: \"e41e2028-117d-4e52-878b-0c722a1891aa\") " pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.300780 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.382995 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 12:44:41 crc kubenswrapper[4809]: W1205 12:44:41.407931 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f939aa0_d8aa_4771_a4fe_326db3a407df.slice/crio-1c111e0621c3d10fb10b518546171bbeba75726a54418d4af83b9c63a7237397 WatchSource:0}: Error finding container 1c111e0621c3d10fb10b518546171bbeba75726a54418d4af83b9c63a7237397: Status 404 returned error can't find the container with id 1c111e0621c3d10fb10b518546171bbeba75726a54418d4af83b9c63a7237397 Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.417591 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.690142 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"2f939aa0-d8aa-4771-a4fe-326db3a407df","Type":"ContainerStarted","Data":"1c111e0621c3d10fb10b518546171bbeba75726a54418d4af83b9c63a7237397"} Dec 05 12:44:41 crc kubenswrapper[4809]: I1205 12:44:41.932598 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 05 12:44:42 crc kubenswrapper[4809]: I1205 12:44:42.707966 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"2f939aa0-d8aa-4771-a4fe-326db3a407df","Type":"ContainerStarted","Data":"bcb03e14e2b657c5780e89f1094fad4090e2053e10856dd60a8d83e9809c06ee"} Dec 05 12:44:42 crc kubenswrapper[4809]: I1205 12:44:42.708384 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"2f939aa0-d8aa-4771-a4fe-326db3a407df","Type":"ContainerStarted","Data":"92465f4c35a8206ad1868a10ffc7f585ff2a7191c75a2e997d2b3d74e0e4784d"} Dec 05 12:44:42 crc kubenswrapper[4809]: I1205 12:44:42.712150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"e41e2028-117d-4e52-878b-0c722a1891aa","Type":"ContainerStarted","Data":"cc1d6f48a3646dd9b2537e4ce421f29eaea9d5cb9e366f1aa08ddb0a0e86d918"} Dec 05 12:44:42 crc kubenswrapper[4809]: I1205 12:44:42.982942 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.81:8776/healthcheck\": dial tcp 10.217.1.81:8776: connect: connection refused" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.552597 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.725882 4809 generic.go:334] "Generic (PLEG): container finished" podID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerID="b9a73c767c2e152f69be560b1810464f231616fceb537886aa5b2f88de292254" exitCode=0 Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.727907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"38ef0e35-a057-43e9-901d-e02be6abcca7","Type":"ContainerDied","Data":"b9a73c767c2e152f69be560b1810464f231616fceb537886aa5b2f88de292254"} Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.844373 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.846248 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.848419 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.872561 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.187147604 podStartE2EDuration="3.872546175s" podCreationTimestamp="2025-12-05 12:44:40 +0000 UTC" firstStartedPulling="2025-12-05 12:44:41.417367332 +0000 UTC m=+5776.808343890" lastFinishedPulling="2025-12-05 12:44:42.102765903 +0000 UTC m=+5777.493742461" observedRunningTime="2025-12-05 12:44:42.741871577 +0000 UTC m=+5778.132848155" watchObservedRunningTime="2025-12-05 12:44:43.872546175 +0000 UTC m=+5779.263522723" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.880716 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.881317 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.882751 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 12:44:43 crc kubenswrapper[4809]: I1205 12:44:43.887789 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.233393 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.400548 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38ef0e35-a057-43e9-901d-e02be6abcca7-etc-machine-id\") pod \"38ef0e35-a057-43e9-901d-e02be6abcca7\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.400620 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data-custom\") pod \"38ef0e35-a057-43e9-901d-e02be6abcca7\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.400679 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ef0e35-a057-43e9-901d-e02be6abcca7-logs\") pod \"38ef0e35-a057-43e9-901d-e02be6abcca7\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.400767 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-combined-ca-bundle\") pod \"38ef0e35-a057-43e9-901d-e02be6abcca7\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.400843 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhvqv\" (UniqueName: \"kubernetes.io/projected/38ef0e35-a057-43e9-901d-e02be6abcca7-kube-api-access-xhvqv\") pod \"38ef0e35-a057-43e9-901d-e02be6abcca7\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.400899 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-scripts\") pod \"38ef0e35-a057-43e9-901d-e02be6abcca7\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.400986 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data\") pod \"38ef0e35-a057-43e9-901d-e02be6abcca7\" (UID: \"38ef0e35-a057-43e9-901d-e02be6abcca7\") " Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.401368 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38ef0e35-a057-43e9-901d-e02be6abcca7-logs" (OuterVolumeSpecName: "logs") pod "38ef0e35-a057-43e9-901d-e02be6abcca7" (UID: "38ef0e35-a057-43e9-901d-e02be6abcca7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.401781 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38ef0e35-a057-43e9-901d-e02be6abcca7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "38ef0e35-a057-43e9-901d-e02be6abcca7" (UID: "38ef0e35-a057-43e9-901d-e02be6abcca7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.409885 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "38ef0e35-a057-43e9-901d-e02be6abcca7" (UID: "38ef0e35-a057-43e9-901d-e02be6abcca7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.409894 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ef0e35-a057-43e9-901d-e02be6abcca7-kube-api-access-xhvqv" (OuterVolumeSpecName: "kube-api-access-xhvqv") pod "38ef0e35-a057-43e9-901d-e02be6abcca7" (UID: "38ef0e35-a057-43e9-901d-e02be6abcca7"). InnerVolumeSpecName "kube-api-access-xhvqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.424782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-scripts" (OuterVolumeSpecName: "scripts") pod "38ef0e35-a057-43e9-901d-e02be6abcca7" (UID: "38ef0e35-a057-43e9-901d-e02be6abcca7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.440060 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38ef0e35-a057-43e9-901d-e02be6abcca7" (UID: "38ef0e35-a057-43e9-901d-e02be6abcca7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.474694 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data" (OuterVolumeSpecName: "config-data") pod "38ef0e35-a057-43e9-901d-e02be6abcca7" (UID: "38ef0e35-a057-43e9-901d-e02be6abcca7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.502996 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38ef0e35-a057-43e9-901d-e02be6abcca7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.503033 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.503043 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ef0e35-a057-43e9-901d-e02be6abcca7-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.503051 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.503061 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhvqv\" (UniqueName: \"kubernetes.io/projected/38ef0e35-a057-43e9-901d-e02be6abcca7-kube-api-access-xhvqv\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.503072 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.503080 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ef0e35-a057-43e9-901d-e02be6abcca7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.739332 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.740424 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"38ef0e35-a057-43e9-901d-e02be6abcca7","Type":"ContainerDied","Data":"0b2a3a8debd7335a0ff5356a67e44a5261e15e8aded8fa0cee8d9d344f76929f"} Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.740523 4809 scope.go:117] "RemoveContainer" containerID="b9a73c767c2e152f69be560b1810464f231616fceb537886aa5b2f88de292254" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.754093 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"e41e2028-117d-4e52-878b-0c722a1891aa","Type":"ContainerStarted","Data":"1fcf8793c63670f0dafa04c4a18d84083c8c32271d8feee2174e0ea882491f15"} Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.754125 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"e41e2028-117d-4e52-878b-0c722a1891aa","Type":"ContainerStarted","Data":"a51de1cae04438bb4e0fcb33e9443aae09a541c0e817657efbfbbda2d79fd1f7"} Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.754140 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.757552 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.763727 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.782488 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=4.108762223 podStartE2EDuration="4.782470969s" podCreationTimestamp="2025-12-05 12:44:40 +0000 UTC" firstStartedPulling="2025-12-05 12:44:41.93067012 +0000 UTC m=+5777.321646678" lastFinishedPulling="2025-12-05 12:44:42.604378866 +0000 UTC m=+5777.995355424" observedRunningTime="2025-12-05 12:44:44.777340051 +0000 UTC m=+5780.168316619" watchObservedRunningTime="2025-12-05 12:44:44.782470969 +0000 UTC m=+5780.173447527" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.783478 4809 scope.go:117] "RemoveContainer" containerID="fbd933c23104016a5dbdbd261a9c6c40593e4cf9ec15ee5cdeaed66e05ae6a7e" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.869473 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.912180 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.912222 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:44 crc kubenswrapper[4809]: E1205 12:44:44.915052 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api-log" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.915082 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api-log" Dec 05 12:44:44 crc kubenswrapper[4809]: E1205 12:44:44.915140 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.915149 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.924410 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.924471 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" containerName="cinder-api-log" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.933664 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.933763 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 12:44:44 crc kubenswrapper[4809]: I1205 12:44:44.939105 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.015736 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d878828-def9-4192-adfc-1479f2e7a252-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.015873 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.015905 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d878828-def9-4192-adfc-1479f2e7a252-logs\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.015983 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-scripts\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.016059 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plvk4\" (UniqueName: \"kubernetes.io/projected/1d878828-def9-4192-adfc-1479f2e7a252-kube-api-access-plvk4\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.016077 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.016108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-config-data\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.117576 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d878828-def9-4192-adfc-1479f2e7a252-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.117861 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.117880 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d878828-def9-4192-adfc-1479f2e7a252-logs\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.117925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-scripts\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.117978 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plvk4\" (UniqueName: \"kubernetes.io/projected/1d878828-def9-4192-adfc-1479f2e7a252-kube-api-access-plvk4\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.117996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.118024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-config-data\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.118928 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d878828-def9-4192-adfc-1479f2e7a252-logs\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.119000 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d878828-def9-4192-adfc-1479f2e7a252-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.123940 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.125294 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-config-data\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.126154 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.132853 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d878828-def9-4192-adfc-1479f2e7a252-scripts\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.172756 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plvk4\" (UniqueName: \"kubernetes.io/projected/1d878828-def9-4192-adfc-1479f2e7a252-kube-api-access-plvk4\") pod \"cinder-api-0\" (UID: \"1d878828-def9-4192-adfc-1479f2e7a252\") " pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.279112 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.713547 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.750104 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.764658 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d878828-def9-4192-adfc-1479f2e7a252","Type":"ContainerStarted","Data":"16e65e2f12c7b20aa435a5ee94358c581963dda62f2bd586eeda423dd78e51f2"} Dec 05 12:44:45 crc kubenswrapper[4809]: I1205 12:44:45.872672 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:44:45 crc kubenswrapper[4809]: E1205 12:44:45.872977 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:44:46 crc kubenswrapper[4809]: I1205 12:44:46.301517 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 05 12:44:46 crc kubenswrapper[4809]: I1205 12:44:46.776067 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d878828-def9-4192-adfc-1479f2e7a252","Type":"ContainerStarted","Data":"af002a227732de2fec5be597301bb58ac1ee13089e4921ffa7fb5e34ea5d8522"} Dec 05 12:44:46 crc kubenswrapper[4809]: I1205 12:44:46.887429 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ef0e35-a057-43e9-901d-e02be6abcca7" path="/var/lib/kubelet/pods/38ef0e35-a057-43e9-901d-e02be6abcca7/volumes" Dec 05 12:44:47 crc kubenswrapper[4809]: I1205 12:44:47.787735 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d878828-def9-4192-adfc-1479f2e7a252","Type":"ContainerStarted","Data":"15226e0a316b68da7c8d6999e0872069cbd4acedc9792c397451bb1a44c92881"} Dec 05 12:44:47 crc kubenswrapper[4809]: I1205 12:44:47.787904 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 12:44:47 crc kubenswrapper[4809]: I1205 12:44:47.814360 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.814341248 podStartE2EDuration="3.814341248s" podCreationTimestamp="2025-12-05 12:44:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:47.806205769 +0000 UTC m=+5783.197182337" watchObservedRunningTime="2025-12-05 12:44:47.814341248 +0000 UTC m=+5783.205317806" Dec 05 12:44:48 crc kubenswrapper[4809]: I1205 12:44:48.754196 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 12:44:48 crc kubenswrapper[4809]: I1205 12:44:48.810842 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:48 crc kubenswrapper[4809]: I1205 12:44:48.811168 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="cinder-scheduler" containerID="cri-o://0063b9770b72d2c09db64a01e4b016123ebea9d1d0b2902a96a0a48be6eee09c" gracePeriod=30 Dec 05 12:44:48 crc kubenswrapper[4809]: I1205 12:44:48.811346 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="probe" containerID="cri-o://d5f26fb137cfe1d96eb5ba1df0c71f711eff14633bc199cad440c9cb3fd07d39" gracePeriod=30 Dec 05 12:44:49 crc kubenswrapper[4809]: I1205 12:44:49.809960 4809 generic.go:334] "Generic (PLEG): container finished" podID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerID="d5f26fb137cfe1d96eb5ba1df0c71f711eff14633bc199cad440c9cb3fd07d39" exitCode=0 Dec 05 12:44:49 crc kubenswrapper[4809]: I1205 12:44:49.810095 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e11bde65-d1a0-46a0-897a-573299b21ad1","Type":"ContainerDied","Data":"d5f26fb137cfe1d96eb5ba1df0c71f711eff14633bc199cad440c9cb3fd07d39"} Dec 05 12:44:50 crc kubenswrapper[4809]: I1205 12:44:50.920408 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 05 12:44:51 crc kubenswrapper[4809]: I1205 12:44:51.557441 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 05 12:44:51 crc kubenswrapper[4809]: I1205 12:44:51.833800 4809 generic.go:334] "Generic (PLEG): container finished" podID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerID="0063b9770b72d2c09db64a01e4b016123ebea9d1d0b2902a96a0a48be6eee09c" exitCode=0 Dec 05 12:44:51 crc kubenswrapper[4809]: I1205 12:44:51.833844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e11bde65-d1a0-46a0-897a-573299b21ad1","Type":"ContainerDied","Data":"0063b9770b72d2c09db64a01e4b016123ebea9d1d0b2902a96a0a48be6eee09c"} Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.304356 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.367557 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data\") pod \"e11bde65-d1a0-46a0-897a-573299b21ad1\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.367668 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-combined-ca-bundle\") pod \"e11bde65-d1a0-46a0-897a-573299b21ad1\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.367823 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-scripts\") pod \"e11bde65-d1a0-46a0-897a-573299b21ad1\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.368045 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data-custom\") pod \"e11bde65-d1a0-46a0-897a-573299b21ad1\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.368117 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg48w\" (UniqueName: \"kubernetes.io/projected/e11bde65-d1a0-46a0-897a-573299b21ad1-kube-api-access-cg48w\") pod \"e11bde65-d1a0-46a0-897a-573299b21ad1\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.368290 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e11bde65-d1a0-46a0-897a-573299b21ad1-etc-machine-id\") pod \"e11bde65-d1a0-46a0-897a-573299b21ad1\" (UID: \"e11bde65-d1a0-46a0-897a-573299b21ad1\") " Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.369298 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e11bde65-d1a0-46a0-897a-573299b21ad1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e11bde65-d1a0-46a0-897a-573299b21ad1" (UID: "e11bde65-d1a0-46a0-897a-573299b21ad1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.374414 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e11bde65-d1a0-46a0-897a-573299b21ad1" (UID: "e11bde65-d1a0-46a0-897a-573299b21ad1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.376739 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e11bde65-d1a0-46a0-897a-573299b21ad1-kube-api-access-cg48w" (OuterVolumeSpecName: "kube-api-access-cg48w") pod "e11bde65-d1a0-46a0-897a-573299b21ad1" (UID: "e11bde65-d1a0-46a0-897a-573299b21ad1"). InnerVolumeSpecName "kube-api-access-cg48w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.381744 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-scripts" (OuterVolumeSpecName: "scripts") pod "e11bde65-d1a0-46a0-897a-573299b21ad1" (UID: "e11bde65-d1a0-46a0-897a-573299b21ad1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.422231 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e11bde65-d1a0-46a0-897a-573299b21ad1" (UID: "e11bde65-d1a0-46a0-897a-573299b21ad1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.471170 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.471422 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg48w\" (UniqueName: \"kubernetes.io/projected/e11bde65-d1a0-46a0-897a-573299b21ad1-kube-api-access-cg48w\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.471526 4809 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e11bde65-d1a0-46a0-897a-573299b21ad1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.471622 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.471758 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.477728 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data" (OuterVolumeSpecName: "config-data") pod "e11bde65-d1a0-46a0-897a-573299b21ad1" (UID: "e11bde65-d1a0-46a0-897a-573299b21ad1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.573428 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bde65-d1a0-46a0-897a-573299b21ad1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.852165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e11bde65-d1a0-46a0-897a-573299b21ad1","Type":"ContainerDied","Data":"28473b30787b940b8e3ffd5fa29923f45c30b40e2b5e44e67f47bd977bdc7d13"} Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.852575 4809 scope.go:117] "RemoveContainer" containerID="d5f26fb137cfe1d96eb5ba1df0c71f711eff14633bc199cad440c9cb3fd07d39" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.852249 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.896471 4809 scope.go:117] "RemoveContainer" containerID="0063b9770b72d2c09db64a01e4b016123ebea9d1d0b2902a96a0a48be6eee09c" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.901204 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.920466 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.933689 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:52 crc kubenswrapper[4809]: E1205 12:44:52.934567 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="probe" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.934618 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="probe" Dec 05 12:44:52 crc kubenswrapper[4809]: E1205 12:44:52.934709 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="cinder-scheduler" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.934723 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="cinder-scheduler" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.935111 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="probe" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.935269 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" containerName="cinder-scheduler" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.936517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.939078 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 12:44:52 crc kubenswrapper[4809]: I1205 12:44:52.953515 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.087906 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.087995 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fjmz\" (UniqueName: \"kubernetes.io/projected/b255ea1a-622c-4d9e-bf97-977a4f359ed5-kube-api-access-7fjmz\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.088068 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-scripts\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.088155 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b255ea1a-622c-4d9e-bf97-977a4f359ed5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.088232 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-config-data\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.088283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.090954 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.091010 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="080276e3-3e86-4d71-a8fc-c5d4927e7031" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.189891 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.189990 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fjmz\" (UniqueName: \"kubernetes.io/projected/b255ea1a-622c-4d9e-bf97-977a4f359ed5-kube-api-access-7fjmz\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.190039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-scripts\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.190114 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b255ea1a-622c-4d9e-bf97-977a4f359ed5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.190286 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b255ea1a-622c-4d9e-bf97-977a4f359ed5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.190428 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-config-data\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.191306 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.195484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.195732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.195813 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-scripts\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.196725 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b255ea1a-622c-4d9e-bf97-977a4f359ed5-config-data\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.207107 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fjmz\" (UniqueName: \"kubernetes.io/projected/b255ea1a-622c-4d9e-bf97-977a4f359ed5-kube-api-access-7fjmz\") pod \"cinder-scheduler-0\" (UID: \"b255ea1a-622c-4d9e-bf97-977a4f359ed5\") " pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.294728 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.741178 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 12:44:53 crc kubenswrapper[4809]: I1205 12:44:53.861908 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b255ea1a-622c-4d9e-bf97-977a4f359ed5","Type":"ContainerStarted","Data":"c3dfa8ab68dfc150e760bcde2435ee7cf1c6aa3ae70edae7e6bc2d0e6567a88f"} Dec 05 12:44:54 crc kubenswrapper[4809]: I1205 12:44:54.892959 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e11bde65-d1a0-46a0-897a-573299b21ad1" path="/var/lib/kubelet/pods/e11bde65-d1a0-46a0-897a-573299b21ad1/volumes" Dec 05 12:44:54 crc kubenswrapper[4809]: I1205 12:44:54.895263 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b255ea1a-622c-4d9e-bf97-977a4f359ed5","Type":"ContainerStarted","Data":"b1881d98e24d3bdecbeb2f9a31258f870fe604b3928b3f82d12baa0d907f0a06"} Dec 05 12:44:54 crc kubenswrapper[4809]: I1205 12:44:54.895374 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b255ea1a-622c-4d9e-bf97-977a4f359ed5","Type":"ContainerStarted","Data":"c27504acc68557c076e2fed875b08111d51a9a19a334b99069cea23ae725e14b"} Dec 05 12:44:54 crc kubenswrapper[4809]: I1205 12:44:54.948306 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.948283122 podStartE2EDuration="2.948283122s" podCreationTimestamp="2025-12-05 12:44:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:44:54.946392421 +0000 UTC m=+5790.337368979" watchObservedRunningTime="2025-12-05 12:44:54.948283122 +0000 UTC m=+5790.339259680" Dec 05 12:44:56 crc kubenswrapper[4809]: I1205 12:44:56.873313 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:44:56 crc kubenswrapper[4809]: E1205 12:44:56.874100 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:44:57 crc kubenswrapper[4809]: I1205 12:44:57.150303 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 12:44:58 crc kubenswrapper[4809]: I1205 12:44:58.295626 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.150021 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss"] Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.151712 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.160421 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.160925 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.163687 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss"] Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.278788 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8912b56b-2f35-41b9-aa28-dba7a16846af-secret-volume\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.278866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrm6\" (UniqueName: \"kubernetes.io/projected/8912b56b-2f35-41b9-aa28-dba7a16846af-kube-api-access-zxrm6\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.279078 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8912b56b-2f35-41b9-aa28-dba7a16846af-config-volume\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.381601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8912b56b-2f35-41b9-aa28-dba7a16846af-secret-volume\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.381688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxrm6\" (UniqueName: \"kubernetes.io/projected/8912b56b-2f35-41b9-aa28-dba7a16846af-kube-api-access-zxrm6\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.381764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8912b56b-2f35-41b9-aa28-dba7a16846af-config-volume\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.382751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8912b56b-2f35-41b9-aa28-dba7a16846af-config-volume\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.388153 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8912b56b-2f35-41b9-aa28-dba7a16846af-secret-volume\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.402930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxrm6\" (UniqueName: \"kubernetes.io/projected/8912b56b-2f35-41b9-aa28-dba7a16846af-kube-api-access-zxrm6\") pod \"collect-profiles-29415645-d42ss\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.482486 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:00 crc kubenswrapper[4809]: W1205 12:45:00.943100 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8912b56b_2f35_41b9_aa28_dba7a16846af.slice/crio-4a44d1ad263faf3a66200922d0cdb861244e5a49886ef57a3e34c9258d87a2d5 WatchSource:0}: Error finding container 4a44d1ad263faf3a66200922d0cdb861244e5a49886ef57a3e34c9258d87a2d5: Status 404 returned error can't find the container with id 4a44d1ad263faf3a66200922d0cdb861244e5a49886ef57a3e34c9258d87a2d5 Dec 05 12:45:00 crc kubenswrapper[4809]: I1205 12:45:00.951246 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss"] Dec 05 12:45:01 crc kubenswrapper[4809]: I1205 12:45:01.968441 4809 generic.go:334] "Generic (PLEG): container finished" podID="8912b56b-2f35-41b9-aa28-dba7a16846af" containerID="ed3732aaa161829f7f1f0932a5fc022983f990af667f0a0a2d6883e075b39bb0" exitCode=0 Dec 05 12:45:01 crc kubenswrapper[4809]: I1205 12:45:01.968485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" event={"ID":"8912b56b-2f35-41b9-aa28-dba7a16846af","Type":"ContainerDied","Data":"ed3732aaa161829f7f1f0932a5fc022983f990af667f0a0a2d6883e075b39bb0"} Dec 05 12:45:01 crc kubenswrapper[4809]: I1205 12:45:01.968885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" event={"ID":"8912b56b-2f35-41b9-aa28-dba7a16846af","Type":"ContainerStarted","Data":"4a44d1ad263faf3a66200922d0cdb861244e5a49886ef57a3e34c9258d87a2d5"} Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.339552 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.445156 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxrm6\" (UniqueName: \"kubernetes.io/projected/8912b56b-2f35-41b9-aa28-dba7a16846af-kube-api-access-zxrm6\") pod \"8912b56b-2f35-41b9-aa28-dba7a16846af\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.445740 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8912b56b-2f35-41b9-aa28-dba7a16846af-config-volume\") pod \"8912b56b-2f35-41b9-aa28-dba7a16846af\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.445956 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8912b56b-2f35-41b9-aa28-dba7a16846af-secret-volume\") pod \"8912b56b-2f35-41b9-aa28-dba7a16846af\" (UID: \"8912b56b-2f35-41b9-aa28-dba7a16846af\") " Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.446506 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8912b56b-2f35-41b9-aa28-dba7a16846af-config-volume" (OuterVolumeSpecName: "config-volume") pod "8912b56b-2f35-41b9-aa28-dba7a16846af" (UID: "8912b56b-2f35-41b9-aa28-dba7a16846af"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.451737 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8912b56b-2f35-41b9-aa28-dba7a16846af-kube-api-access-zxrm6" (OuterVolumeSpecName: "kube-api-access-zxrm6") pod "8912b56b-2f35-41b9-aa28-dba7a16846af" (UID: "8912b56b-2f35-41b9-aa28-dba7a16846af"). InnerVolumeSpecName "kube-api-access-zxrm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.452416 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8912b56b-2f35-41b9-aa28-dba7a16846af-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8912b56b-2f35-41b9-aa28-dba7a16846af" (UID: "8912b56b-2f35-41b9-aa28-dba7a16846af"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.519455 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.547926 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8912b56b-2f35-41b9-aa28-dba7a16846af-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.548167 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8912b56b-2f35-41b9-aa28-dba7a16846af-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.548226 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxrm6\" (UniqueName: \"kubernetes.io/projected/8912b56b-2f35-41b9-aa28-dba7a16846af-kube-api-access-zxrm6\") on node \"crc\" DevicePath \"\"" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.986887 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" event={"ID":"8912b56b-2f35-41b9-aa28-dba7a16846af","Type":"ContainerDied","Data":"4a44d1ad263faf3a66200922d0cdb861244e5a49886ef57a3e34c9258d87a2d5"} Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.987186 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a44d1ad263faf3a66200922d0cdb861244e5a49886ef57a3e34c9258d87a2d5" Dec 05 12:45:03 crc kubenswrapper[4809]: I1205 12:45:03.986940 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss" Dec 05 12:45:04 crc kubenswrapper[4809]: I1205 12:45:04.420220 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr"] Dec 05 12:45:04 crc kubenswrapper[4809]: I1205 12:45:04.428112 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415600-kv5lr"] Dec 05 12:45:04 crc kubenswrapper[4809]: I1205 12:45:04.902884 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daa47f26-d043-470e-a6d9-030daba81545" path="/var/lib/kubelet/pods/daa47f26-d043-470e-a6d9-030daba81545/volumes" Dec 05 12:45:08 crc kubenswrapper[4809]: I1205 12:45:08.872489 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:45:08 crc kubenswrapper[4809]: E1205 12:45:08.873377 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:45:19 crc kubenswrapper[4809]: I1205 12:45:19.871954 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:45:19 crc kubenswrapper[4809]: E1205 12:45:19.872601 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:45:31 crc kubenswrapper[4809]: I1205 12:45:31.871947 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:45:31 crc kubenswrapper[4809]: E1205 12:45:31.872776 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:45:35 crc kubenswrapper[4809]: I1205 12:45:35.382258 4809 scope.go:117] "RemoveContainer" containerID="f336e1d363bde633a5587d6a97ae454e2a99c824f9f765c91e0f445d4872b5f8" Dec 05 12:45:35 crc kubenswrapper[4809]: I1205 12:45:35.402481 4809 scope.go:117] "RemoveContainer" containerID="fe39dec510a18d985b99fdac8a723904afc0b00035aa86aa9c60c37824a06055" Dec 05 12:45:35 crc kubenswrapper[4809]: I1205 12:45:35.423595 4809 scope.go:117] "RemoveContainer" containerID="09171c2d9eccca7f147ed9f35aafa857991474b6a5d157f5eb9154cf262b7af7" Dec 05 12:45:43 crc kubenswrapper[4809]: I1205 12:45:43.872671 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:45:43 crc kubenswrapper[4809]: E1205 12:45:43.873732 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:45:55 crc kubenswrapper[4809]: I1205 12:45:55.872396 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:45:55 crc kubenswrapper[4809]: E1205 12:45:55.873329 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:46:10 crc kubenswrapper[4809]: I1205 12:46:10.872880 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:46:10 crc kubenswrapper[4809]: E1205 12:46:10.873782 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:46:22 crc kubenswrapper[4809]: I1205 12:46:22.873428 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:46:22 crc kubenswrapper[4809]: E1205 12:46:22.874113 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:46:27 crc kubenswrapper[4809]: I1205 12:46:27.051809 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-7gk2d"] Dec 05 12:46:27 crc kubenswrapper[4809]: I1205 12:46:27.062946 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-93ae-account-create-update-hsvmz"] Dec 05 12:46:27 crc kubenswrapper[4809]: I1205 12:46:27.073686 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-7gk2d"] Dec 05 12:46:27 crc kubenswrapper[4809]: I1205 12:46:27.084555 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-93ae-account-create-update-hsvmz"] Dec 05 12:46:28 crc kubenswrapper[4809]: I1205 12:46:28.908565 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5afa0578-156c-428c-84f0-066a27a4b928" path="/var/lib/kubelet/pods/5afa0578-156c-428c-84f0-066a27a4b928/volumes" Dec 05 12:46:28 crc kubenswrapper[4809]: I1205 12:46:28.909407 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8" path="/var/lib/kubelet/pods/a7b82e35-91ca-4f6d-af5d-4f7ed343d5d8/volumes" Dec 05 12:46:35 crc kubenswrapper[4809]: I1205 12:46:35.049272 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-t7kj2"] Dec 05 12:46:35 crc kubenswrapper[4809]: I1205 12:46:35.057901 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-t7kj2"] Dec 05 12:46:35 crc kubenswrapper[4809]: I1205 12:46:35.546726 4809 scope.go:117] "RemoveContainer" containerID="320a423d0e9842ec6d36250de676aa392b41e43300fc8a512d63d217ae9323ea" Dec 05 12:46:35 crc kubenswrapper[4809]: I1205 12:46:35.568929 4809 scope.go:117] "RemoveContainer" containerID="0cd26242653c78c8a97098fa6b28c64b5ba6eabd8a0a124f658c653f324c804d" Dec 05 12:46:35 crc kubenswrapper[4809]: I1205 12:46:35.618683 4809 scope.go:117] "RemoveContainer" containerID="2c2828fc46bb16840eb4ac349d1b41a6d84ca0e8d8173fc8f214555df95cca8d" Dec 05 12:46:35 crc kubenswrapper[4809]: I1205 12:46:35.692942 4809 scope.go:117] "RemoveContainer" containerID="c35995a52f9855a7ef88701b2d863ea4e611683bbda69f5579ae01dd654124bc" Dec 05 12:46:35 crc kubenswrapper[4809]: I1205 12:46:35.872598 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:46:35 crc kubenswrapper[4809]: E1205 12:46:35.872944 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:46:36 crc kubenswrapper[4809]: I1205 12:46:36.884711 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9aea549-b837-4a23-bdcb-570f7a9edaab" path="/var/lib/kubelet/pods/a9aea549-b837-4a23-bdcb-570f7a9edaab/volumes" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.195598 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hm2jg"] Dec 05 12:46:42 crc kubenswrapper[4809]: E1205 12:46:42.196826 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8912b56b-2f35-41b9-aa28-dba7a16846af" containerName="collect-profiles" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.196842 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8912b56b-2f35-41b9-aa28-dba7a16846af" containerName="collect-profiles" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.197043 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8912b56b-2f35-41b9-aa28-dba7a16846af" containerName="collect-profiles" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.197836 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.201721 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-8fnrb" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.202549 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.225206 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hm2jg"] Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.247730 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-run-ovn\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.247798 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpf5v\" (UniqueName: \"kubernetes.io/projected/60ff38f8-1c69-42c3-8968-9276b9af9dce-kube-api-access-lpf5v\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.247851 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60ff38f8-1c69-42c3-8968-9276b9af9dce-scripts\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.248000 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-run\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.248060 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-log-ovn\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.261532 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-sfspc"] Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.263712 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.278983 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sfspc"] Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.349978 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-etc-ovs\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350032 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-run\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350109 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-log\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-run\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350196 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-log-ovn\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/273a55f4-4590-47d6-8428-69f365f8ee54-scripts\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350276 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-run-ovn\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350295 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4n5k\" (UniqueName: \"kubernetes.io/projected/273a55f4-4590-47d6-8428-69f365f8ee54-kube-api-access-m4n5k\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350333 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpf5v\" (UniqueName: \"kubernetes.io/projected/60ff38f8-1c69-42c3-8968-9276b9af9dce-kube-api-access-lpf5v\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350378 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60ff38f8-1c69-42c3-8968-9276b9af9dce-scripts\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350454 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-lib\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-run-ovn\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-run\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.350990 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/60ff38f8-1c69-42c3-8968-9276b9af9dce-var-log-ovn\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.353055 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60ff38f8-1c69-42c3-8968-9276b9af9dce-scripts\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.373874 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpf5v\" (UniqueName: \"kubernetes.io/projected/60ff38f8-1c69-42c3-8968-9276b9af9dce-kube-api-access-lpf5v\") pod \"ovn-controller-hm2jg\" (UID: \"60ff38f8-1c69-42c3-8968-9276b9af9dce\") " pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.452542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-lib\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.452615 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-etc-ovs\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.452790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-run\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.452823 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-log\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.452879 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/273a55f4-4590-47d6-8428-69f365f8ee54-scripts\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.452898 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4n5k\" (UniqueName: \"kubernetes.io/projected/273a55f4-4590-47d6-8428-69f365f8ee54-kube-api-access-m4n5k\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.453292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-run\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.453352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-etc-ovs\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.453352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-lib\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.453352 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/273a55f4-4590-47d6-8428-69f365f8ee54-var-log\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.455314 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/273a55f4-4590-47d6-8428-69f365f8ee54-scripts\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.476237 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4n5k\" (UniqueName: \"kubernetes.io/projected/273a55f4-4590-47d6-8428-69f365f8ee54-kube-api-access-m4n5k\") pod \"ovn-controller-ovs-sfspc\" (UID: \"273a55f4-4590-47d6-8428-69f365f8ee54\") " pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.530370 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:42 crc kubenswrapper[4809]: I1205 12:46:42.583410 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:43 crc kubenswrapper[4809]: I1205 12:46:43.190307 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hm2jg"] Dec 05 12:46:43 crc kubenswrapper[4809]: I1205 12:46:43.576326 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sfspc"] Dec 05 12:46:43 crc kubenswrapper[4809]: W1205 12:46:43.584902 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod273a55f4_4590_47d6_8428_69f365f8ee54.slice/crio-b5db795d8492c507e5beee1e538665793db8bb44c5e266f8e54471baed1c71f9 WatchSource:0}: Error finding container b5db795d8492c507e5beee1e538665793db8bb44c5e266f8e54471baed1c71f9: Status 404 returned error can't find the container with id b5db795d8492c507e5beee1e538665793db8bb44c5e266f8e54471baed1c71f9 Dec 05 12:46:43 crc kubenswrapper[4809]: I1205 12:46:43.870359 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-khns6"] Dec 05 12:46:43 crc kubenswrapper[4809]: I1205 12:46:43.871983 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:43 crc kubenswrapper[4809]: I1205 12:46:43.882099 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-khns6"] Dec 05 12:46:43 crc kubenswrapper[4809]: I1205 12:46:43.909563 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.001664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sfspc" event={"ID":"273a55f4-4590-47d6-8428-69f365f8ee54","Type":"ContainerStarted","Data":"4a981f8a70c1db0e47ca0275205616a49ea1932e715264326b9ddbc98dab1827"} Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.001712 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sfspc" event={"ID":"273a55f4-4590-47d6-8428-69f365f8ee54","Type":"ContainerStarted","Data":"b5db795d8492c507e5beee1e538665793db8bb44c5e266f8e54471baed1c71f9"} Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.004909 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hm2jg" event={"ID":"60ff38f8-1c69-42c3-8968-9276b9af9dce","Type":"ContainerStarted","Data":"38b5d1e64c80b21a1469dd1755165cc7331997515df29d16b78e727025af4d14"} Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.004951 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hm2jg" event={"ID":"60ff38f8-1c69-42c3-8968-9276b9af9dce","Type":"ContainerStarted","Data":"fb94b9a9159bffab60c21c9c2a0f274e0c80e309957724a81519512c4f9e3ad9"} Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.005733 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-hm2jg" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.014170 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/40aadb59-e0c4-4370-9ce0-6899c84771ee-ovs-rundir\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.014279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40aadb59-e0c4-4370-9ce0-6899c84771ee-config\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.014314 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67fnv\" (UniqueName: \"kubernetes.io/projected/40aadb59-e0c4-4370-9ce0-6899c84771ee-kube-api-access-67fnv\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.014920 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/40aadb59-e0c4-4370-9ce0-6899c84771ee-ovn-rundir\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.064152 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-hm2jg" podStartSLOduration=2.064130409 podStartE2EDuration="2.064130409s" podCreationTimestamp="2025-12-05 12:46:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:46:44.051581931 +0000 UTC m=+5899.442558499" watchObservedRunningTime="2025-12-05 12:46:44.064130409 +0000 UTC m=+5899.455106977" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.117370 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/40aadb59-e0c4-4370-9ce0-6899c84771ee-ovs-rundir\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.117617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40aadb59-e0c4-4370-9ce0-6899c84771ee-config\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.117750 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67fnv\" (UniqueName: \"kubernetes.io/projected/40aadb59-e0c4-4370-9ce0-6899c84771ee-kube-api-access-67fnv\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.117877 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/40aadb59-e0c4-4370-9ce0-6899c84771ee-ovn-rundir\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.119372 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/40aadb59-e0c4-4370-9ce0-6899c84771ee-ovs-rundir\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.119511 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/40aadb59-e0c4-4370-9ce0-6899c84771ee-ovn-rundir\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.120172 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40aadb59-e0c4-4370-9ce0-6899c84771ee-config\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.139218 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67fnv\" (UniqueName: \"kubernetes.io/projected/40aadb59-e0c4-4370-9ce0-6899c84771ee-kube-api-access-67fnv\") pod \"ovn-controller-metrics-khns6\" (UID: \"40aadb59-e0c4-4370-9ce0-6899c84771ee\") " pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.232176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-khns6" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.239620 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-ljckw"] Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.241067 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.258437 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ljckw"] Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.322120 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t8wp\" (UniqueName: \"kubernetes.io/projected/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-kube-api-access-7t8wp\") pod \"octavia-db-create-ljckw\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.322317 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-operator-scripts\") pod \"octavia-db-create-ljckw\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.424562 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-operator-scripts\") pod \"octavia-db-create-ljckw\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.425051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t8wp\" (UniqueName: \"kubernetes.io/projected/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-kube-api-access-7t8wp\") pod \"octavia-db-create-ljckw\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.425590 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-operator-scripts\") pod \"octavia-db-create-ljckw\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.453898 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t8wp\" (UniqueName: \"kubernetes.io/projected/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-kube-api-access-7t8wp\") pod \"octavia-db-create-ljckw\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.734272 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:44 crc kubenswrapper[4809]: I1205 12:46:44.743615 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-khns6"] Dec 05 12:46:45 crc kubenswrapper[4809]: I1205 12:46:45.022613 4809 generic.go:334] "Generic (PLEG): container finished" podID="273a55f4-4590-47d6-8428-69f365f8ee54" containerID="4a981f8a70c1db0e47ca0275205616a49ea1932e715264326b9ddbc98dab1827" exitCode=0 Dec 05 12:46:45 crc kubenswrapper[4809]: I1205 12:46:45.022894 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sfspc" event={"ID":"273a55f4-4590-47d6-8428-69f365f8ee54","Type":"ContainerDied","Data":"4a981f8a70c1db0e47ca0275205616a49ea1932e715264326b9ddbc98dab1827"} Dec 05 12:46:45 crc kubenswrapper[4809]: I1205 12:46:45.028585 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-khns6" event={"ID":"40aadb59-e0c4-4370-9ce0-6899c84771ee","Type":"ContainerStarted","Data":"eed3316d69762d43e6c13fa522d11f5adc29b7add6c6ae3789560e47008eda31"} Dec 05 12:46:45 crc kubenswrapper[4809]: I1205 12:46:45.247858 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ljckw"] Dec 05 12:46:45 crc kubenswrapper[4809]: W1205 12:46:45.264012 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cd5acc7_eaa1_407a_ac38_c6d904b8e710.slice/crio-4f3cb8b85bd89c8ad167851890a82a799039af1d7d12f47ee9fcbcc4d3d03333 WatchSource:0}: Error finding container 4f3cb8b85bd89c8ad167851890a82a799039af1d7d12f47ee9fcbcc4d3d03333: Status 404 returned error can't find the container with id 4f3cb8b85bd89c8ad167851890a82a799039af1d7d12f47ee9fcbcc4d3d03333 Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.043256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ljckw" event={"ID":"7cd5acc7-eaa1-407a-ac38-c6d904b8e710","Type":"ContainerStarted","Data":"118479dc83bbf3f90a720c0acbf80435a293c0194d6dcb7a9a2271743aac5774"} Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.043862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ljckw" event={"ID":"7cd5acc7-eaa1-407a-ac38-c6d904b8e710","Type":"ContainerStarted","Data":"4f3cb8b85bd89c8ad167851890a82a799039af1d7d12f47ee9fcbcc4d3d03333"} Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.051491 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-khns6" event={"ID":"40aadb59-e0c4-4370-9ce0-6899c84771ee","Type":"ContainerStarted","Data":"8707adc499225965966c7bbe7c36c0eb516c533cb5381f079aebcd8525f5c6a9"} Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.056482 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sfspc" event={"ID":"273a55f4-4590-47d6-8428-69f365f8ee54","Type":"ContainerStarted","Data":"ef80831c85d20458a63000d170ba649c744c5da47dbb00c20ba48a2a3e7d6b26"} Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.056519 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sfspc" event={"ID":"273a55f4-4590-47d6-8428-69f365f8ee54","Type":"ContainerStarted","Data":"23ab96bd232844e7aef79b337dee18dfb916efc6268f1477ff5608606f072fd5"} Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.056554 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.056575 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.070499 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-create-ljckw" podStartSLOduration=2.07048185 podStartE2EDuration="2.07048185s" podCreationTimestamp="2025-12-05 12:46:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:46:46.062765482 +0000 UTC m=+5901.453742040" watchObservedRunningTime="2025-12-05 12:46:46.07048185 +0000 UTC m=+5901.461458408" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.083677 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-khns6" podStartSLOduration=3.083653594 podStartE2EDuration="3.083653594s" podCreationTimestamp="2025-12-05 12:46:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:46:46.076149382 +0000 UTC m=+5901.467125960" watchObservedRunningTime="2025-12-05 12:46:46.083653594 +0000 UTC m=+5901.474630152" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.111708 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-sfspc" podStartSLOduration=4.111689308 podStartE2EDuration="4.111689308s" podCreationTimestamp="2025-12-05 12:46:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:46:46.109179341 +0000 UTC m=+5901.500155909" watchObservedRunningTime="2025-12-05 12:46:46.111689308 +0000 UTC m=+5901.502665866" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.258479 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-8dfb-account-create-update-kw22n"] Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.259894 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.261948 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.270420 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-8dfb-account-create-update-kw22n"] Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.378881 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h4kp\" (UniqueName: \"kubernetes.io/projected/b825193a-8c7b-45c6-80cd-e929b8771e25-kube-api-access-4h4kp\") pod \"octavia-8dfb-account-create-update-kw22n\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.379034 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b825193a-8c7b-45c6-80cd-e929b8771e25-operator-scripts\") pod \"octavia-8dfb-account-create-update-kw22n\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.480557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b825193a-8c7b-45c6-80cd-e929b8771e25-operator-scripts\") pod \"octavia-8dfb-account-create-update-kw22n\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.480732 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h4kp\" (UniqueName: \"kubernetes.io/projected/b825193a-8c7b-45c6-80cd-e929b8771e25-kube-api-access-4h4kp\") pod \"octavia-8dfb-account-create-update-kw22n\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.481572 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b825193a-8c7b-45c6-80cd-e929b8771e25-operator-scripts\") pod \"octavia-8dfb-account-create-update-kw22n\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.513799 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h4kp\" (UniqueName: \"kubernetes.io/projected/b825193a-8c7b-45c6-80cd-e929b8771e25-kube-api-access-4h4kp\") pod \"octavia-8dfb-account-create-update-kw22n\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:46 crc kubenswrapper[4809]: I1205 12:46:46.575002 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:47 crc kubenswrapper[4809]: I1205 12:46:47.069861 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-8dfb-account-create-update-kw22n"] Dec 05 12:46:47 crc kubenswrapper[4809]: I1205 12:46:47.071884 4809 generic.go:334] "Generic (PLEG): container finished" podID="7cd5acc7-eaa1-407a-ac38-c6d904b8e710" containerID="118479dc83bbf3f90a720c0acbf80435a293c0194d6dcb7a9a2271743aac5774" exitCode=0 Dec 05 12:46:47 crc kubenswrapper[4809]: I1205 12:46:47.072025 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ljckw" event={"ID":"7cd5acc7-eaa1-407a-ac38-c6d904b8e710","Type":"ContainerDied","Data":"118479dc83bbf3f90a720c0acbf80435a293c0194d6dcb7a9a2271743aac5774"} Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.082589 4809 generic.go:334] "Generic (PLEG): container finished" podID="b825193a-8c7b-45c6-80cd-e929b8771e25" containerID="069cf1f28b90d6d086c50a6997d1575bc89bcbac8de589f870361c9759ded0c3" exitCode=0 Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.082741 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-8dfb-account-create-update-kw22n" event={"ID":"b825193a-8c7b-45c6-80cd-e929b8771e25","Type":"ContainerDied","Data":"069cf1f28b90d6d086c50a6997d1575bc89bcbac8de589f870361c9759ded0c3"} Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.083020 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-8dfb-account-create-update-kw22n" event={"ID":"b825193a-8c7b-45c6-80cd-e929b8771e25","Type":"ContainerStarted","Data":"9d44a0c0170c79329bbcf35c81047f0e19600590896368b9f859c0bf816eb3ed"} Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.426146 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.521404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-operator-scripts\") pod \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.521559 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t8wp\" (UniqueName: \"kubernetes.io/projected/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-kube-api-access-7t8wp\") pod \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\" (UID: \"7cd5acc7-eaa1-407a-ac38-c6d904b8e710\") " Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.522229 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7cd5acc7-eaa1-407a-ac38-c6d904b8e710" (UID: "7cd5acc7-eaa1-407a-ac38-c6d904b8e710"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.526678 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-kube-api-access-7t8wp" (OuterVolumeSpecName: "kube-api-access-7t8wp") pod "7cd5acc7-eaa1-407a-ac38-c6d904b8e710" (UID: "7cd5acc7-eaa1-407a-ac38-c6d904b8e710"). InnerVolumeSpecName "kube-api-access-7t8wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.623550 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.623595 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t8wp\" (UniqueName: \"kubernetes.io/projected/7cd5acc7-eaa1-407a-ac38-c6d904b8e710-kube-api-access-7t8wp\") on node \"crc\" DevicePath \"\"" Dec 05 12:46:48 crc kubenswrapper[4809]: I1205 12:46:48.871988 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:46:48 crc kubenswrapper[4809]: E1205 12:46:48.872359 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.051564 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bhwls"] Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.080422 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bhwls"] Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.093881 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ljckw" Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.093873 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ljckw" event={"ID":"7cd5acc7-eaa1-407a-ac38-c6d904b8e710","Type":"ContainerDied","Data":"4f3cb8b85bd89c8ad167851890a82a799039af1d7d12f47ee9fcbcc4d3d03333"} Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.094066 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f3cb8b85bd89c8ad167851890a82a799039af1d7d12f47ee9fcbcc4d3d03333" Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.446019 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.544669 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b825193a-8c7b-45c6-80cd-e929b8771e25-operator-scripts\") pod \"b825193a-8c7b-45c6-80cd-e929b8771e25\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.544825 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h4kp\" (UniqueName: \"kubernetes.io/projected/b825193a-8c7b-45c6-80cd-e929b8771e25-kube-api-access-4h4kp\") pod \"b825193a-8c7b-45c6-80cd-e929b8771e25\" (UID: \"b825193a-8c7b-45c6-80cd-e929b8771e25\") " Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.545738 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b825193a-8c7b-45c6-80cd-e929b8771e25-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b825193a-8c7b-45c6-80cd-e929b8771e25" (UID: "b825193a-8c7b-45c6-80cd-e929b8771e25"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.555962 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b825193a-8c7b-45c6-80cd-e929b8771e25-kube-api-access-4h4kp" (OuterVolumeSpecName: "kube-api-access-4h4kp") pod "b825193a-8c7b-45c6-80cd-e929b8771e25" (UID: "b825193a-8c7b-45c6-80cd-e929b8771e25"). InnerVolumeSpecName "kube-api-access-4h4kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.647485 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h4kp\" (UniqueName: \"kubernetes.io/projected/b825193a-8c7b-45c6-80cd-e929b8771e25-kube-api-access-4h4kp\") on node \"crc\" DevicePath \"\"" Dec 05 12:46:49 crc kubenswrapper[4809]: I1205 12:46:49.647529 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b825193a-8c7b-45c6-80cd-e929b8771e25-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.104036 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-8dfb-account-create-update-kw22n" event={"ID":"b825193a-8c7b-45c6-80cd-e929b8771e25","Type":"ContainerDied","Data":"9d44a0c0170c79329bbcf35c81047f0e19600590896368b9f859c0bf816eb3ed"} Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.104351 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d44a0c0170c79329bbcf35c81047f0e19600590896368b9f859c0bf816eb3ed" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.104287 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-8dfb-account-create-update-kw22n" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.579164 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6jzpp"] Dec 05 12:46:50 crc kubenswrapper[4809]: E1205 12:46:50.579541 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd5acc7-eaa1-407a-ac38-c6d904b8e710" containerName="mariadb-database-create" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.579557 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd5acc7-eaa1-407a-ac38-c6d904b8e710" containerName="mariadb-database-create" Dec 05 12:46:50 crc kubenswrapper[4809]: E1205 12:46:50.579590 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b825193a-8c7b-45c6-80cd-e929b8771e25" containerName="mariadb-account-create-update" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.579596 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b825193a-8c7b-45c6-80cd-e929b8771e25" containerName="mariadb-account-create-update" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.579794 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b825193a-8c7b-45c6-80cd-e929b8771e25" containerName="mariadb-account-create-update" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.579809 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd5acc7-eaa1-407a-ac38-c6d904b8e710" containerName="mariadb-database-create" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.581105 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.591025 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jzpp"] Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.664044 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf9hw\" (UniqueName: \"kubernetes.io/projected/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-kube-api-access-rf9hw\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.664124 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-utilities\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.664153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-catalog-content\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.766343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf9hw\" (UniqueName: \"kubernetes.io/projected/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-kube-api-access-rf9hw\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.766451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-utilities\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.766471 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-catalog-content\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.767174 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-catalog-content\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.767243 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-utilities\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.783713 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf9hw\" (UniqueName: \"kubernetes.io/projected/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-kube-api-access-rf9hw\") pod \"community-operators-6jzpp\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.884599 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea6d817-5ff8-4e2f-b39d-b61187f79ec8" path="/var/lib/kubelet/pods/dea6d817-5ff8-4e2f-b39d-b61187f79ec8/volumes" Dec 05 12:46:50 crc kubenswrapper[4809]: I1205 12:46:50.899688 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:46:51 crc kubenswrapper[4809]: I1205 12:46:51.408635 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jzpp"] Dec 05 12:46:51 crc kubenswrapper[4809]: W1205 12:46:51.414186 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf6c7c90_58eb_428b_af7b_45bc7d9d9fd5.slice/crio-a3543539088ff5c29f57efd9cfd89c1836317e6bfba500bdc46efb606f00f45b WatchSource:0}: Error finding container a3543539088ff5c29f57efd9cfd89c1836317e6bfba500bdc46efb606f00f45b: Status 404 returned error can't find the container with id a3543539088ff5c29f57efd9cfd89c1836317e6bfba500bdc46efb606f00f45b Dec 05 12:46:52 crc kubenswrapper[4809]: I1205 12:46:52.128295 4809 generic.go:334] "Generic (PLEG): container finished" podID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerID="cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb" exitCode=0 Dec 05 12:46:52 crc kubenswrapper[4809]: I1205 12:46:52.128405 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jzpp" event={"ID":"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5","Type":"ContainerDied","Data":"cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb"} Dec 05 12:46:52 crc kubenswrapper[4809]: I1205 12:46:52.128711 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jzpp" event={"ID":"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5","Type":"ContainerStarted","Data":"a3543539088ff5c29f57efd9cfd89c1836317e6bfba500bdc46efb606f00f45b"} Dec 05 12:46:53 crc kubenswrapper[4809]: I1205 12:46:53.141622 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jzpp" event={"ID":"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5","Type":"ContainerStarted","Data":"0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8"} Dec 05 12:46:54 crc kubenswrapper[4809]: I1205 12:46:54.160327 4809 generic.go:334] "Generic (PLEG): container finished" podID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerID="0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8" exitCode=0 Dec 05 12:46:54 crc kubenswrapper[4809]: I1205 12:46:54.160590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jzpp" event={"ID":"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5","Type":"ContainerDied","Data":"0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8"} Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.171166 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jzpp" event={"ID":"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5","Type":"ContainerStarted","Data":"2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10"} Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.193293 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6jzpp" podStartSLOduration=2.478328982 podStartE2EDuration="5.19327752s" podCreationTimestamp="2025-12-05 12:46:50 +0000 UTC" firstStartedPulling="2025-12-05 12:46:52.129994458 +0000 UTC m=+5907.520971016" lastFinishedPulling="2025-12-05 12:46:54.844942996 +0000 UTC m=+5910.235919554" observedRunningTime="2025-12-05 12:46:55.187541935 +0000 UTC m=+5910.578518513" watchObservedRunningTime="2025-12-05 12:46:55.19327752 +0000 UTC m=+5910.584254078" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.237174 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-q9566"] Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.238456 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.247197 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-q9566"] Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.350046 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eba5b4d-1157-4cb8-982d-aca3713a3311-operator-scripts\") pod \"octavia-persistence-db-create-q9566\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.350501 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2lvt\" (UniqueName: \"kubernetes.io/projected/9eba5b4d-1157-4cb8-982d-aca3713a3311-kube-api-access-g2lvt\") pod \"octavia-persistence-db-create-q9566\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.452254 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2lvt\" (UniqueName: \"kubernetes.io/projected/9eba5b4d-1157-4cb8-982d-aca3713a3311-kube-api-access-g2lvt\") pod \"octavia-persistence-db-create-q9566\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.452403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eba5b4d-1157-4cb8-982d-aca3713a3311-operator-scripts\") pod \"octavia-persistence-db-create-q9566\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.453192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eba5b4d-1157-4cb8-982d-aca3713a3311-operator-scripts\") pod \"octavia-persistence-db-create-q9566\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.494259 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2lvt\" (UniqueName: \"kubernetes.io/projected/9eba5b4d-1157-4cb8-982d-aca3713a3311-kube-api-access-g2lvt\") pod \"octavia-persistence-db-create-q9566\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:55 crc kubenswrapper[4809]: I1205 12:46:55.555094 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.078206 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-q9566"] Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.181061 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-q9566" event={"ID":"9eba5b4d-1157-4cb8-982d-aca3713a3311","Type":"ContainerStarted","Data":"4291b691534acf74a532cdecf6baf03a31a30ccb5f6752f7ef1f1ce002b3b71a"} Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.263948 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-2e9e-account-create-update-qh759"] Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.265311 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.267241 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.271621 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0a6148f-dfac-4671-a743-e16dbf29a4d4-operator-scripts\") pod \"octavia-2e9e-account-create-update-qh759\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.272260 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfbw5\" (UniqueName: \"kubernetes.io/projected/b0a6148f-dfac-4671-a743-e16dbf29a4d4-kube-api-access-sfbw5\") pod \"octavia-2e9e-account-create-update-qh759\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.291017 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-2e9e-account-create-update-qh759"] Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.374067 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0a6148f-dfac-4671-a743-e16dbf29a4d4-operator-scripts\") pod \"octavia-2e9e-account-create-update-qh759\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.374259 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfbw5\" (UniqueName: \"kubernetes.io/projected/b0a6148f-dfac-4671-a743-e16dbf29a4d4-kube-api-access-sfbw5\") pod \"octavia-2e9e-account-create-update-qh759\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.375408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0a6148f-dfac-4671-a743-e16dbf29a4d4-operator-scripts\") pod \"octavia-2e9e-account-create-update-qh759\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.392769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfbw5\" (UniqueName: \"kubernetes.io/projected/b0a6148f-dfac-4671-a743-e16dbf29a4d4-kube-api-access-sfbw5\") pod \"octavia-2e9e-account-create-update-qh759\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:56 crc kubenswrapper[4809]: I1205 12:46:56.599738 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:57 crc kubenswrapper[4809]: W1205 12:46:57.152030 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0a6148f_dfac_4671_a743_e16dbf29a4d4.slice/crio-755dcc7cb14edc263049336af476e4869d7c951c9212ec94fd0472fe3fcd7fb7 WatchSource:0}: Error finding container 755dcc7cb14edc263049336af476e4869d7c951c9212ec94fd0472fe3fcd7fb7: Status 404 returned error can't find the container with id 755dcc7cb14edc263049336af476e4869d7c951c9212ec94fd0472fe3fcd7fb7 Dec 05 12:46:57 crc kubenswrapper[4809]: I1205 12:46:57.160479 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-2e9e-account-create-update-qh759"] Dec 05 12:46:57 crc kubenswrapper[4809]: I1205 12:46:57.193993 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-2e9e-account-create-update-qh759" event={"ID":"b0a6148f-dfac-4671-a743-e16dbf29a4d4","Type":"ContainerStarted","Data":"755dcc7cb14edc263049336af476e4869d7c951c9212ec94fd0472fe3fcd7fb7"} Dec 05 12:46:57 crc kubenswrapper[4809]: I1205 12:46:57.198496 4809 generic.go:334] "Generic (PLEG): container finished" podID="9eba5b4d-1157-4cb8-982d-aca3713a3311" containerID="b9b3a0f6ee5c8638652650e608983e6164eb5f2c5cf7ea33f2a61b7f7833d350" exitCode=0 Dec 05 12:46:57 crc kubenswrapper[4809]: I1205 12:46:57.198523 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-q9566" event={"ID":"9eba5b4d-1157-4cb8-982d-aca3713a3311","Type":"ContainerDied","Data":"b9b3a0f6ee5c8638652650e608983e6164eb5f2c5cf7ea33f2a61b7f7833d350"} Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.219159 4809 generic.go:334] "Generic (PLEG): container finished" podID="b0a6148f-dfac-4671-a743-e16dbf29a4d4" containerID="fa17c5af786b0d2b9fda3f78e0f50517d6419286baf9ca775ce3d0ee186807d3" exitCode=0 Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.219291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-2e9e-account-create-update-qh759" event={"ID":"b0a6148f-dfac-4671-a743-e16dbf29a4d4","Type":"ContainerDied","Data":"fa17c5af786b0d2b9fda3f78e0f50517d6419286baf9ca775ce3d0ee186807d3"} Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.623539 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.752404 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eba5b4d-1157-4cb8-982d-aca3713a3311-operator-scripts\") pod \"9eba5b4d-1157-4cb8-982d-aca3713a3311\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.752529 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2lvt\" (UniqueName: \"kubernetes.io/projected/9eba5b4d-1157-4cb8-982d-aca3713a3311-kube-api-access-g2lvt\") pod \"9eba5b4d-1157-4cb8-982d-aca3713a3311\" (UID: \"9eba5b4d-1157-4cb8-982d-aca3713a3311\") " Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.753203 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eba5b4d-1157-4cb8-982d-aca3713a3311-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9eba5b4d-1157-4cb8-982d-aca3713a3311" (UID: "9eba5b4d-1157-4cb8-982d-aca3713a3311"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.757906 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eba5b4d-1157-4cb8-982d-aca3713a3311-kube-api-access-g2lvt" (OuterVolumeSpecName: "kube-api-access-g2lvt") pod "9eba5b4d-1157-4cb8-982d-aca3713a3311" (UID: "9eba5b4d-1157-4cb8-982d-aca3713a3311"). InnerVolumeSpecName "kube-api-access-g2lvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.855844 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2lvt\" (UniqueName: \"kubernetes.io/projected/9eba5b4d-1157-4cb8-982d-aca3713a3311-kube-api-access-g2lvt\") on node \"crc\" DevicePath \"\"" Dec 05 12:46:58 crc kubenswrapper[4809]: I1205 12:46:58.856255 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eba5b4d-1157-4cb8-982d-aca3713a3311-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:46:59 crc kubenswrapper[4809]: E1205 12:46:59.073791 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eba5b4d_1157_4cb8_982d_aca3713a3311.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eba5b4d_1157_4cb8_982d_aca3713a3311.slice/crio-4291b691534acf74a532cdecf6baf03a31a30ccb5f6752f7ef1f1ce002b3b71a\": RecentStats: unable to find data in memory cache]" Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.230147 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-q9566" event={"ID":"9eba5b4d-1157-4cb8-982d-aca3713a3311","Type":"ContainerDied","Data":"4291b691534acf74a532cdecf6baf03a31a30ccb5f6752f7ef1f1ce002b3b71a"} Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.230202 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4291b691534acf74a532cdecf6baf03a31a30ccb5f6752f7ef1f1ce002b3b71a" Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.230166 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-q9566" Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.626329 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.668221 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0a6148f-dfac-4671-a743-e16dbf29a4d4-operator-scripts\") pod \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.668310 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfbw5\" (UniqueName: \"kubernetes.io/projected/b0a6148f-dfac-4671-a743-e16dbf29a4d4-kube-api-access-sfbw5\") pod \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\" (UID: \"b0a6148f-dfac-4671-a743-e16dbf29a4d4\") " Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.669982 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a6148f-dfac-4671-a743-e16dbf29a4d4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0a6148f-dfac-4671-a743-e16dbf29a4d4" (UID: "b0a6148f-dfac-4671-a743-e16dbf29a4d4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.694731 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0a6148f-dfac-4671-a743-e16dbf29a4d4-kube-api-access-sfbw5" (OuterVolumeSpecName: "kube-api-access-sfbw5") pod "b0a6148f-dfac-4671-a743-e16dbf29a4d4" (UID: "b0a6148f-dfac-4671-a743-e16dbf29a4d4"). InnerVolumeSpecName "kube-api-access-sfbw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.771115 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0a6148f-dfac-4671-a743-e16dbf29a4d4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:46:59 crc kubenswrapper[4809]: I1205 12:46:59.771148 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfbw5\" (UniqueName: \"kubernetes.io/projected/b0a6148f-dfac-4671-a743-e16dbf29a4d4-kube-api-access-sfbw5\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:00 crc kubenswrapper[4809]: I1205 12:47:00.241150 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-2e9e-account-create-update-qh759" event={"ID":"b0a6148f-dfac-4671-a743-e16dbf29a4d4","Type":"ContainerDied","Data":"755dcc7cb14edc263049336af476e4869d7c951c9212ec94fd0472fe3fcd7fb7"} Dec 05 12:47:00 crc kubenswrapper[4809]: I1205 12:47:00.241487 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="755dcc7cb14edc263049336af476e4869d7c951c9212ec94fd0472fe3fcd7fb7" Dec 05 12:47:00 crc kubenswrapper[4809]: I1205 12:47:00.241197 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-2e9e-account-create-update-qh759" Dec 05 12:47:00 crc kubenswrapper[4809]: I1205 12:47:00.900285 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:47:00 crc kubenswrapper[4809]: I1205 12:47:00.901956 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:47:01 crc kubenswrapper[4809]: I1205 12:47:01.029594 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:47:01 crc kubenswrapper[4809]: I1205 12:47:01.300256 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:47:01 crc kubenswrapper[4809]: I1205 12:47:01.872590 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:47:01 crc kubenswrapper[4809]: E1205 12:47:01.872975 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.461507 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-59bccf98fd-m9c2j"] Dec 05 12:47:02 crc kubenswrapper[4809]: E1205 12:47:02.461933 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eba5b4d-1157-4cb8-982d-aca3713a3311" containerName="mariadb-database-create" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.461946 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eba5b4d-1157-4cb8-982d-aca3713a3311" containerName="mariadb-database-create" Dec 05 12:47:02 crc kubenswrapper[4809]: E1205 12:47:02.461966 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a6148f-dfac-4671-a743-e16dbf29a4d4" containerName="mariadb-account-create-update" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.461973 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a6148f-dfac-4671-a743-e16dbf29a4d4" containerName="mariadb-account-create-update" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.462243 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0a6148f-dfac-4671-a743-e16dbf29a4d4" containerName="mariadb-account-create-update" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.462261 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eba5b4d-1157-4cb8-982d-aca3713a3311" containerName="mariadb-database-create" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.463750 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.465896 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.466039 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-c9tbh" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.467476 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.480286 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-59bccf98fd-m9c2j"] Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.531339 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-scripts\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.531460 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/6fd18514-6933-428e-95eb-de00e63a5acb-octavia-run\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.531489 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6fd18514-6933-428e-95eb-de00e63a5acb-config-data-merged\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.531719 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-combined-ca-bundle\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.531868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-config-data\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.633470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-scripts\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.633560 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/6fd18514-6933-428e-95eb-de00e63a5acb-octavia-run\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.633601 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6fd18514-6933-428e-95eb-de00e63a5acb-config-data-merged\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.633694 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-combined-ca-bundle\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.633773 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-config-data\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.636206 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6fd18514-6933-428e-95eb-de00e63a5acb-config-data-merged\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.636206 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/6fd18514-6933-428e-95eb-de00e63a5acb-octavia-run\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.643026 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-config-data\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.647959 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-combined-ca-bundle\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.651440 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fd18514-6933-428e-95eb-de00e63a5acb-scripts\") pod \"octavia-api-59bccf98fd-m9c2j\" (UID: \"6fd18514-6933-428e-95eb-de00e63a5acb\") " pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:02 crc kubenswrapper[4809]: I1205 12:47:02.803132 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:03 crc kubenswrapper[4809]: I1205 12:47:03.316001 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-59bccf98fd-m9c2j"] Dec 05 12:47:04 crc kubenswrapper[4809]: I1205 12:47:04.284934 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59bccf98fd-m9c2j" event={"ID":"6fd18514-6933-428e-95eb-de00e63a5acb","Type":"ContainerStarted","Data":"00a86c9cd54eeaea6f080580e53f62f7a4b32c582f18a8cd78762e467d2526d4"} Dec 05 12:47:04 crc kubenswrapper[4809]: I1205 12:47:04.558177 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jzpp"] Dec 05 12:47:04 crc kubenswrapper[4809]: I1205 12:47:04.558433 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6jzpp" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="registry-server" containerID="cri-o://2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10" gracePeriod=2 Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.128572 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.209794 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-utilities\") pod \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.209917 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf9hw\" (UniqueName: \"kubernetes.io/projected/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-kube-api-access-rf9hw\") pod \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.210126 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-catalog-content\") pod \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\" (UID: \"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5\") " Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.213150 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-utilities" (OuterVolumeSpecName: "utilities") pod "bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" (UID: "bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.219976 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-kube-api-access-rf9hw" (OuterVolumeSpecName: "kube-api-access-rf9hw") pod "bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" (UID: "bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5"). InnerVolumeSpecName "kube-api-access-rf9hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.282993 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" (UID: "bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.312721 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf9hw\" (UniqueName: \"kubernetes.io/projected/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-kube-api-access-rf9hw\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.312765 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.312775 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.344844 4809 generic.go:334] "Generic (PLEG): container finished" podID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerID="2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10" exitCode=0 Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.344891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jzpp" event={"ID":"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5","Type":"ContainerDied","Data":"2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10"} Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.344918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jzpp" event={"ID":"bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5","Type":"ContainerDied","Data":"a3543539088ff5c29f57efd9cfd89c1836317e6bfba500bdc46efb606f00f45b"} Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.344936 4809 scope.go:117] "RemoveContainer" containerID="2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.345088 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jzpp" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.403713 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jzpp"] Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.417806 4809 scope.go:117] "RemoveContainer" containerID="0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.432186 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6jzpp"] Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.467589 4809 scope.go:117] "RemoveContainer" containerID="cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.549215 4809 scope.go:117] "RemoveContainer" containerID="2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10" Dec 05 12:47:05 crc kubenswrapper[4809]: E1205 12:47:05.549718 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10\": container with ID starting with 2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10 not found: ID does not exist" containerID="2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.549764 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10"} err="failed to get container status \"2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10\": rpc error: code = NotFound desc = could not find container \"2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10\": container with ID starting with 2f565f3b9b3f110de3ef48947919fe8845a0efde45936809537890d35f31cf10 not found: ID does not exist" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.549784 4809 scope.go:117] "RemoveContainer" containerID="0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8" Dec 05 12:47:05 crc kubenswrapper[4809]: E1205 12:47:05.550025 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8\": container with ID starting with 0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8 not found: ID does not exist" containerID="0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.550043 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8"} err="failed to get container status \"0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8\": rpc error: code = NotFound desc = could not find container \"0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8\": container with ID starting with 0c93993c2ed586a2f255ae5e1eb973fad680a0c16987410d5fdd4164abb152a8 not found: ID does not exist" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.550154 4809 scope.go:117] "RemoveContainer" containerID="cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb" Dec 05 12:47:05 crc kubenswrapper[4809]: E1205 12:47:05.550575 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb\": container with ID starting with cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb not found: ID does not exist" containerID="cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb" Dec 05 12:47:05 crc kubenswrapper[4809]: I1205 12:47:05.550606 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb"} err="failed to get container status \"cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb\": rpc error: code = NotFound desc = could not find container \"cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb\": container with ID starting with cd254afa848634c3dde32c584d3d27679b923329949d98c858c67b2701d57dfb not found: ID does not exist" Dec 05 12:47:06 crc kubenswrapper[4809]: I1205 12:47:06.888406 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" path="/var/lib/kubelet/pods/bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5/volumes" Dec 05 12:47:14 crc kubenswrapper[4809]: I1205 12:47:14.431453 4809 generic.go:334] "Generic (PLEG): container finished" podID="6fd18514-6933-428e-95eb-de00e63a5acb" containerID="62b6abd5ee9fb48f0387de1d11f45acac11ead317ddb0396eaf99583f4147aec" exitCode=0 Dec 05 12:47:14 crc kubenswrapper[4809]: I1205 12:47:14.431512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59bccf98fd-m9c2j" event={"ID":"6fd18514-6933-428e-95eb-de00e63a5acb","Type":"ContainerDied","Data":"62b6abd5ee9fb48f0387de1d11f45acac11ead317ddb0396eaf99583f4147aec"} Dec 05 12:47:15 crc kubenswrapper[4809]: I1205 12:47:15.444798 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59bccf98fd-m9c2j" event={"ID":"6fd18514-6933-428e-95eb-de00e63a5acb","Type":"ContainerStarted","Data":"d1de0b3463e69e98c59d9e69870b90359cdc0312bd3b7a09c145bf136bdc6c44"} Dec 05 12:47:15 crc kubenswrapper[4809]: I1205 12:47:15.445110 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-59bccf98fd-m9c2j" event={"ID":"6fd18514-6933-428e-95eb-de00e63a5acb","Type":"ContainerStarted","Data":"096adabb7e61631d2c311e5b29e5c2db40dc1f389da61cd56e628312285f43c6"} Dec 05 12:47:15 crc kubenswrapper[4809]: I1205 12:47:15.445276 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:15 crc kubenswrapper[4809]: I1205 12:47:15.470545 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-59bccf98fd-m9c2j" podStartSLOduration=2.9331651880000003 podStartE2EDuration="13.470522465s" podCreationTimestamp="2025-12-05 12:47:02 +0000 UTC" firstStartedPulling="2025-12-05 12:47:03.330702786 +0000 UTC m=+5918.721679334" lastFinishedPulling="2025-12-05 12:47:13.868060053 +0000 UTC m=+5929.259036611" observedRunningTime="2025-12-05 12:47:15.462989382 +0000 UTC m=+5930.853965970" watchObservedRunningTime="2025-12-05 12:47:15.470522465 +0000 UTC m=+5930.861499033" Dec 05 12:47:15 crc kubenswrapper[4809]: I1205 12:47:15.875914 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:47:15 crc kubenswrapper[4809]: E1205 12:47:15.876224 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:47:16 crc kubenswrapper[4809]: I1205 12:47:16.452398 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.580499 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hm2jg" podUID="60ff38f8-1c69-42c3-8968-9276b9af9dce" containerName="ovn-controller" probeResult="failure" output=< Dec 05 12:47:17 crc kubenswrapper[4809]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 12:47:17 crc kubenswrapper[4809]: > Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.628818 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.630837 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sfspc" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.762479 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hm2jg-config-jvwwv"] Dec 05 12:47:17 crc kubenswrapper[4809]: E1205 12:47:17.763490 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="registry-server" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.763514 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="registry-server" Dec 05 12:47:17 crc kubenswrapper[4809]: E1205 12:47:17.763546 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="extract-utilities" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.763554 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="extract-utilities" Dec 05 12:47:17 crc kubenswrapper[4809]: E1205 12:47:17.763575 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="extract-content" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.763602 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="extract-content" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.764298 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf6c7c90-58eb-428b-af7b-45bc7d9d9fd5" containerName="registry-server" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.764967 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.766774 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.775363 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hm2jg-config-jvwwv"] Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.894564 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-log-ovn\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.897379 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-scripts\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.897454 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk2jv\" (UniqueName: \"kubernetes.io/projected/eb3c7ef2-4002-4c19-b62f-d0856211e629-kube-api-access-pk2jv\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.897520 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.897562 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-additional-scripts\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:17 crc kubenswrapper[4809]: I1205 12:47:17.897624 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run-ovn\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.009461 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-scripts\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.010058 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk2jv\" (UniqueName: \"kubernetes.io/projected/eb3c7ef2-4002-4c19-b62f-d0856211e629-kube-api-access-pk2jv\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.010491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.010723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-additional-scripts\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.010935 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run-ovn\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.011101 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.011266 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run-ovn\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.011517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-log-ovn\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.012341 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-additional-scripts\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.012857 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-log-ovn\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.013230 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-scripts\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.038373 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk2jv\" (UniqueName: \"kubernetes.io/projected/eb3c7ef2-4002-4c19-b62f-d0856211e629-kube-api-access-pk2jv\") pod \"ovn-controller-hm2jg-config-jvwwv\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.128363 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:18 crc kubenswrapper[4809]: W1205 12:47:18.628340 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb3c7ef2_4002_4c19_b62f_d0856211e629.slice/crio-c17347a408323b1a5850edb02231522c038eb9655ebea041e0e8104d53fe662b WatchSource:0}: Error finding container c17347a408323b1a5850edb02231522c038eb9655ebea041e0e8104d53fe662b: Status 404 returned error can't find the container with id c17347a408323b1a5850edb02231522c038eb9655ebea041e0e8104d53fe662b Dec 05 12:47:18 crc kubenswrapper[4809]: I1205 12:47:18.636950 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hm2jg-config-jvwwv"] Dec 05 12:47:19 crc kubenswrapper[4809]: I1205 12:47:19.493270 4809 generic.go:334] "Generic (PLEG): container finished" podID="eb3c7ef2-4002-4c19-b62f-d0856211e629" containerID="c39dc3f714c42f6c61812c5362fbb6e3bfded98b503aa52a548067b44d180c1e" exitCode=0 Dec 05 12:47:19 crc kubenswrapper[4809]: I1205 12:47:19.493329 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hm2jg-config-jvwwv" event={"ID":"eb3c7ef2-4002-4c19-b62f-d0856211e629","Type":"ContainerDied","Data":"c39dc3f714c42f6c61812c5362fbb6e3bfded98b503aa52a548067b44d180c1e"} Dec 05 12:47:19 crc kubenswrapper[4809]: I1205 12:47:19.493582 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hm2jg-config-jvwwv" event={"ID":"eb3c7ef2-4002-4c19-b62f-d0856211e629","Type":"ContainerStarted","Data":"c17347a408323b1a5850edb02231522c038eb9655ebea041e0e8104d53fe662b"} Dec 05 12:47:20 crc kubenswrapper[4809]: I1205 12:47:20.899421 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.080756 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk2jv\" (UniqueName: \"kubernetes.io/projected/eb3c7ef2-4002-4c19-b62f-d0856211e629-kube-api-access-pk2jv\") pod \"eb3c7ef2-4002-4c19-b62f-d0856211e629\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.080870 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-scripts\") pod \"eb3c7ef2-4002-4c19-b62f-d0856211e629\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081001 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-log-ovn\") pod \"eb3c7ef2-4002-4c19-b62f-d0856211e629\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081031 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run\") pod \"eb3c7ef2-4002-4c19-b62f-d0856211e629\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run-ovn\") pod \"eb3c7ef2-4002-4c19-b62f-d0856211e629\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081126 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-additional-scripts\") pod \"eb3c7ef2-4002-4c19-b62f-d0856211e629\" (UID: \"eb3c7ef2-4002-4c19-b62f-d0856211e629\") " Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "eb3c7ef2-4002-4c19-b62f-d0856211e629" (UID: "eb3c7ef2-4002-4c19-b62f-d0856211e629"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081244 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run" (OuterVolumeSpecName: "var-run") pod "eb3c7ef2-4002-4c19-b62f-d0856211e629" (UID: "eb3c7ef2-4002-4c19-b62f-d0856211e629"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081268 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "eb3c7ef2-4002-4c19-b62f-d0856211e629" (UID: "eb3c7ef2-4002-4c19-b62f-d0856211e629"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081520 4809 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081538 4809 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081549 4809 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb3c7ef2-4002-4c19-b62f-d0856211e629-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "eb3c7ef2-4002-4c19-b62f-d0856211e629" (UID: "eb3c7ef2-4002-4c19-b62f-d0856211e629"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.081941 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-scripts" (OuterVolumeSpecName: "scripts") pod "eb3c7ef2-4002-4c19-b62f-d0856211e629" (UID: "eb3c7ef2-4002-4c19-b62f-d0856211e629"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.086741 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb3c7ef2-4002-4c19-b62f-d0856211e629-kube-api-access-pk2jv" (OuterVolumeSpecName: "kube-api-access-pk2jv") pod "eb3c7ef2-4002-4c19-b62f-d0856211e629" (UID: "eb3c7ef2-4002-4c19-b62f-d0856211e629"). InnerVolumeSpecName "kube-api-access-pk2jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.183188 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk2jv\" (UniqueName: \"kubernetes.io/projected/eb3c7ef2-4002-4c19-b62f-d0856211e629-kube-api-access-pk2jv\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.183223 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.183235 4809 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eb3c7ef2-4002-4c19-b62f-d0856211e629-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.513971 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hm2jg-config-jvwwv" event={"ID":"eb3c7ef2-4002-4c19-b62f-d0856211e629","Type":"ContainerDied","Data":"c17347a408323b1a5850edb02231522c038eb9655ebea041e0e8104d53fe662b"} Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.514016 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hm2jg-config-jvwwv" Dec 05 12:47:21 crc kubenswrapper[4809]: I1205 12:47:21.514019 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c17347a408323b1a5850edb02231522c038eb9655ebea041e0e8104d53fe662b" Dec 05 12:47:22 crc kubenswrapper[4809]: I1205 12:47:22.006373 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hm2jg-config-jvwwv"] Dec 05 12:47:22 crc kubenswrapper[4809]: I1205 12:47:22.019167 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hm2jg-config-jvwwv"] Dec 05 12:47:22 crc kubenswrapper[4809]: I1205 12:47:22.589287 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-hm2jg" Dec 05 12:47:22 crc kubenswrapper[4809]: I1205 12:47:22.883424 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb3c7ef2-4002-4c19-b62f-d0856211e629" path="/var/lib/kubelet/pods/eb3c7ef2-4002-4c19-b62f-d0856211e629/volumes" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.171435 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-64hxq"] Dec 05 12:47:24 crc kubenswrapper[4809]: E1205 12:47:24.172362 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb3c7ef2-4002-4c19-b62f-d0856211e629" containerName="ovn-config" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.172381 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb3c7ef2-4002-4c19-b62f-d0856211e629" containerName="ovn-config" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.172615 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb3c7ef2-4002-4c19-b62f-d0856211e629" containerName="ovn-config" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.174325 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.177113 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.177175 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.177451 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.183161 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-64hxq"] Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.245421 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d8b98d1c-896e-49cc-aead-f30b8afbf33c-config-data-merged\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.245519 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8b98d1c-896e-49cc-aead-f30b8afbf33c-scripts\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.245552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b98d1c-896e-49cc-aead-f30b8afbf33c-config-data\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.245577 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d8b98d1c-896e-49cc-aead-f30b8afbf33c-hm-ports\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.347705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d8b98d1c-896e-49cc-aead-f30b8afbf33c-config-data-merged\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.347756 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8b98d1c-896e-49cc-aead-f30b8afbf33c-scripts\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.347782 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b98d1c-896e-49cc-aead-f30b8afbf33c-config-data\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.347803 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d8b98d1c-896e-49cc-aead-f30b8afbf33c-hm-ports\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.348155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d8b98d1c-896e-49cc-aead-f30b8afbf33c-config-data-merged\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.348742 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d8b98d1c-896e-49cc-aead-f30b8afbf33c-hm-ports\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.353078 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8b98d1c-896e-49cc-aead-f30b8afbf33c-scripts\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.354732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8b98d1c-896e-49cc-aead-f30b8afbf33c-config-data\") pod \"octavia-rsyslog-64hxq\" (UID: \"d8b98d1c-896e-49cc-aead-f30b8afbf33c\") " pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.502727 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.840447 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-kpf5q"] Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.842853 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.845186 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.850303 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-kpf5q"] Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.859336 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/121dd8c3-c175-4312-8d3c-4d3509908e7d-amphora-image\") pod \"octavia-image-upload-59f8cff499-kpf5q\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.859442 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/121dd8c3-c175-4312-8d3c-4d3509908e7d-httpd-config\") pod \"octavia-image-upload-59f8cff499-kpf5q\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.960943 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/121dd8c3-c175-4312-8d3c-4d3509908e7d-httpd-config\") pod \"octavia-image-upload-59f8cff499-kpf5q\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.961319 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/121dd8c3-c175-4312-8d3c-4d3509908e7d-amphora-image\") pod \"octavia-image-upload-59f8cff499-kpf5q\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.961744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/121dd8c3-c175-4312-8d3c-4d3509908e7d-amphora-image\") pod \"octavia-image-upload-59f8cff499-kpf5q\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.965788 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 05 12:47:24 crc kubenswrapper[4809]: I1205 12:47:24.984594 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/121dd8c3-c175-4312-8d3c-4d3509908e7d-httpd-config\") pod \"octavia-image-upload-59f8cff499-kpf5q\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:25 crc kubenswrapper[4809]: I1205 12:47:25.089471 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-64hxq"] Dec 05 12:47:25 crc kubenswrapper[4809]: I1205 12:47:25.183796 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:47:25 crc kubenswrapper[4809]: W1205 12:47:25.558424 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod121dd8c3_c175_4312_8d3c_4d3509908e7d.slice/crio-6b7d847552298ab23041f69db576c62b2d7c0e4034429712ba345b5084b3ddb6 WatchSource:0}: Error finding container 6b7d847552298ab23041f69db576c62b2d7c0e4034429712ba345b5084b3ddb6: Status 404 returned error can't find the container with id 6b7d847552298ab23041f69db576c62b2d7c0e4034429712ba345b5084b3ddb6 Dec 05 12:47:25 crc kubenswrapper[4809]: I1205 12:47:25.559365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-64hxq" event={"ID":"d8b98d1c-896e-49cc-aead-f30b8afbf33c","Type":"ContainerStarted","Data":"3bfac55bda13826a7bdf60c97249e192ebbfa173111d144e20209f4346aac417"} Dec 05 12:47:25 crc kubenswrapper[4809]: I1205 12:47:25.569011 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-kpf5q"] Dec 05 12:47:26 crc kubenswrapper[4809]: I1205 12:47:26.569637 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" event={"ID":"121dd8c3-c175-4312-8d3c-4d3509908e7d","Type":"ContainerStarted","Data":"6b7d847552298ab23041f69db576c62b2d7c0e4034429712ba345b5084b3ddb6"} Dec 05 12:47:26 crc kubenswrapper[4809]: I1205 12:47:26.874230 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:47:26 crc kubenswrapper[4809]: E1205 12:47:26.874852 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:47:27 crc kubenswrapper[4809]: I1205 12:47:27.590398 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-64hxq" event={"ID":"d8b98d1c-896e-49cc-aead-f30b8afbf33c","Type":"ContainerStarted","Data":"90608ace5e80437aec203663592633ba3842e16dadffaf0de3daa0f93929306d"} Dec 05 12:47:29 crc kubenswrapper[4809]: I1205 12:47:29.614397 4809 generic.go:334] "Generic (PLEG): container finished" podID="d8b98d1c-896e-49cc-aead-f30b8afbf33c" containerID="90608ace5e80437aec203663592633ba3842e16dadffaf0de3daa0f93929306d" exitCode=0 Dec 05 12:47:29 crc kubenswrapper[4809]: I1205 12:47:29.614533 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-64hxq" event={"ID":"d8b98d1c-896e-49cc-aead-f30b8afbf33c","Type":"ContainerDied","Data":"90608ace5e80437aec203663592633ba3842e16dadffaf0de3daa0f93929306d"} Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.184924 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-ch655"] Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.187324 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-ch655"] Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.187436 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.198238 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.215511 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-scripts\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.215580 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data-merged\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.215673 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-combined-ca-bundle\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.215732 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.317581 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-scripts\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.317652 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data-merged\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.317707 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-combined-ca-bundle\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.317745 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.318258 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data-merged\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.323883 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-scripts\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.336536 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-combined-ca-bundle\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.347771 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data\") pod \"octavia-db-sync-ch655\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:31 crc kubenswrapper[4809]: I1205 12:47:31.528126 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:32 crc kubenswrapper[4809]: I1205 12:47:32.295267 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-ch655"] Dec 05 12:47:33 crc kubenswrapper[4809]: W1205 12:47:33.420567 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1230fe8_b1b1_443e_adb6_37287dabebb6.slice/crio-c05513e83b251621b2f38e3f525e92ad5c0812b9d2e7635df8c1714f0c60e832 WatchSource:0}: Error finding container c05513e83b251621b2f38e3f525e92ad5c0812b9d2e7635df8c1714f0c60e832: Status 404 returned error can't find the container with id c05513e83b251621b2f38e3f525e92ad5c0812b9d2e7635df8c1714f0c60e832 Dec 05 12:47:33 crc kubenswrapper[4809]: I1205 12:47:33.667714 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ch655" event={"ID":"d1230fe8-b1b1-443e-adb6-37287dabebb6","Type":"ContainerStarted","Data":"c05513e83b251621b2f38e3f525e92ad5c0812b9d2e7635df8c1714f0c60e832"} Dec 05 12:47:35 crc kubenswrapper[4809]: I1205 12:47:35.771447 4809 scope.go:117] "RemoveContainer" containerID="523cb049c5b8d92f1ce557d6bfb614323a3222bea8388b9e4d19d4aabd2e817b" Dec 05 12:47:37 crc kubenswrapper[4809]: I1205 12:47:37.878077 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:37 crc kubenswrapper[4809]: I1205 12:47:37.884084 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-59bccf98fd-m9c2j" Dec 05 12:47:37 crc kubenswrapper[4809]: I1205 12:47:37.998324 4809 scope.go:117] "RemoveContainer" containerID="6c8ae1eb2adf92e436e69ee7ebdb5cac204463a961f97e0b1a9a838bfff47152" Dec 05 12:47:39 crc kubenswrapper[4809]: I1205 12:47:39.738420 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-64hxq" event={"ID":"d8b98d1c-896e-49cc-aead-f30b8afbf33c","Type":"ContainerStarted","Data":"f84e567acaf29e451cead8fca35d9c4d392464c3b72f0222334934ec5b51cb87"} Dec 05 12:47:39 crc kubenswrapper[4809]: I1205 12:47:39.739160 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:47:39 crc kubenswrapper[4809]: I1205 12:47:39.740272 4809 generic.go:334] "Generic (PLEG): container finished" podID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerID="79f8a08a450d1f37503e36ff68824b52b7e792ae9247ef02791daf8e912007b8" exitCode=0 Dec 05 12:47:39 crc kubenswrapper[4809]: I1205 12:47:39.740353 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" event={"ID":"121dd8c3-c175-4312-8d3c-4d3509908e7d","Type":"ContainerDied","Data":"79f8a08a450d1f37503e36ff68824b52b7e792ae9247ef02791daf8e912007b8"} Dec 05 12:47:39 crc kubenswrapper[4809]: I1205 12:47:39.742428 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1230fe8-b1b1-443e-adb6-37287dabebb6" containerID="e8753961dc3469d948ac5d34760c9e083a3055650a0d54ab2419d13aa437d376" exitCode=0 Dec 05 12:47:39 crc kubenswrapper[4809]: I1205 12:47:39.742469 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ch655" event={"ID":"d1230fe8-b1b1-443e-adb6-37287dabebb6","Type":"ContainerDied","Data":"e8753961dc3469d948ac5d34760c9e083a3055650a0d54ab2419d13aa437d376"} Dec 05 12:47:39 crc kubenswrapper[4809]: I1205 12:47:39.768251 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-64hxq" podStartSLOduration=2.401850395 podStartE2EDuration="15.768210717s" podCreationTimestamp="2025-12-05 12:47:24 +0000 UTC" firstStartedPulling="2025-12-05 12:47:25.119713373 +0000 UTC m=+5940.510689931" lastFinishedPulling="2025-12-05 12:47:38.486073695 +0000 UTC m=+5953.877050253" observedRunningTime="2025-12-05 12:47:39.756182183 +0000 UTC m=+5955.147158751" watchObservedRunningTime="2025-12-05 12:47:39.768210717 +0000 UTC m=+5955.159187285" Dec 05 12:47:40 crc kubenswrapper[4809]: I1205 12:47:40.752597 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ch655" event={"ID":"d1230fe8-b1b1-443e-adb6-37287dabebb6","Type":"ContainerStarted","Data":"d53f4578f2ab7748f51f27627f4071ff00be3dc846bf2723ff92133aabce6e5e"} Dec 05 12:47:40 crc kubenswrapper[4809]: I1205 12:47:40.788138 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-ch655" podStartSLOduration=9.788116322 podStartE2EDuration="9.788116322s" podCreationTimestamp="2025-12-05 12:47:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:47:40.767922238 +0000 UTC m=+5956.158898816" watchObservedRunningTime="2025-12-05 12:47:40.788116322 +0000 UTC m=+5956.179092890" Dec 05 12:47:41 crc kubenswrapper[4809]: I1205 12:47:41.768463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" event={"ID":"121dd8c3-c175-4312-8d3c-4d3509908e7d","Type":"ContainerStarted","Data":"916183b43ceb3e0ddda63423c2725edd4e1c6b8f21531521484baa210fcdcebf"} Dec 05 12:47:41 crc kubenswrapper[4809]: I1205 12:47:41.820751 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" podStartSLOduration=2.380718736 podStartE2EDuration="17.820731959s" podCreationTimestamp="2025-12-05 12:47:24 +0000 UTC" firstStartedPulling="2025-12-05 12:47:25.564171213 +0000 UTC m=+5940.955147771" lastFinishedPulling="2025-12-05 12:47:41.004184436 +0000 UTC m=+5956.395160994" observedRunningTime="2025-12-05 12:47:41.782066759 +0000 UTC m=+5957.173043317" watchObservedRunningTime="2025-12-05 12:47:41.820731959 +0000 UTC m=+5957.211708507" Dec 05 12:47:41 crc kubenswrapper[4809]: I1205 12:47:41.872911 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:47:41 crc kubenswrapper[4809]: E1205 12:47:41.873283 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:47:46 crc kubenswrapper[4809]: I1205 12:47:46.828081 4809 generic.go:334] "Generic (PLEG): container finished" podID="d1230fe8-b1b1-443e-adb6-37287dabebb6" containerID="d53f4578f2ab7748f51f27627f4071ff00be3dc846bf2723ff92133aabce6e5e" exitCode=0 Dec 05 12:47:46 crc kubenswrapper[4809]: I1205 12:47:46.828174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ch655" event={"ID":"d1230fe8-b1b1-443e-adb6-37287dabebb6","Type":"ContainerDied","Data":"d53f4578f2ab7748f51f27627f4071ff00be3dc846bf2723ff92133aabce6e5e"} Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.245262 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.277478 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data\") pod \"d1230fe8-b1b1-443e-adb6-37287dabebb6\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.277543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-scripts\") pod \"d1230fe8-b1b1-443e-adb6-37287dabebb6\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.277588 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-combined-ca-bundle\") pod \"d1230fe8-b1b1-443e-adb6-37287dabebb6\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.277667 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data-merged\") pod \"d1230fe8-b1b1-443e-adb6-37287dabebb6\" (UID: \"d1230fe8-b1b1-443e-adb6-37287dabebb6\") " Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.304100 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-scripts" (OuterVolumeSpecName: "scripts") pod "d1230fe8-b1b1-443e-adb6-37287dabebb6" (UID: "d1230fe8-b1b1-443e-adb6-37287dabebb6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.306674 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data" (OuterVolumeSpecName: "config-data") pod "d1230fe8-b1b1-443e-adb6-37287dabebb6" (UID: "d1230fe8-b1b1-443e-adb6-37287dabebb6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.325672 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "d1230fe8-b1b1-443e-adb6-37287dabebb6" (UID: "d1230fe8-b1b1-443e-adb6-37287dabebb6"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.333544 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1230fe8-b1b1-443e-adb6-37287dabebb6" (UID: "d1230fe8-b1b1-443e-adb6-37287dabebb6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.380239 4809 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data-merged\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.380274 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.380283 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.380291 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1230fe8-b1b1-443e-adb6-37287dabebb6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.858295 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ch655" event={"ID":"d1230fe8-b1b1-443e-adb6-37287dabebb6","Type":"ContainerDied","Data":"c05513e83b251621b2f38e3f525e92ad5c0812b9d2e7635df8c1714f0c60e832"} Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.858367 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c05513e83b251621b2f38e3f525e92ad5c0812b9d2e7635df8c1714f0c60e832" Dec 05 12:47:48 crc kubenswrapper[4809]: I1205 12:47:48.858406 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ch655" Dec 05 12:47:52 crc kubenswrapper[4809]: I1205 12:47:52.872314 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:47:52 crc kubenswrapper[4809]: E1205 12:47:52.873009 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:47:54 crc kubenswrapper[4809]: I1205 12:47:54.530315 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-64hxq" Dec 05 12:48:04 crc kubenswrapper[4809]: I1205 12:48:04.878417 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:48:04 crc kubenswrapper[4809]: E1205 12:48:04.879408 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:48:11 crc kubenswrapper[4809]: I1205 12:48:11.943710 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-kpf5q"] Dec 05 12:48:11 crc kubenswrapper[4809]: I1205 12:48:11.944430 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" podUID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerName="octavia-amphora-httpd" containerID="cri-o://916183b43ceb3e0ddda63423c2725edd4e1c6b8f21531521484baa210fcdcebf" gracePeriod=30 Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.068873 4809 generic.go:334] "Generic (PLEG): container finished" podID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerID="916183b43ceb3e0ddda63423c2725edd4e1c6b8f21531521484baa210fcdcebf" exitCode=0 Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.069219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" event={"ID":"121dd8c3-c175-4312-8d3c-4d3509908e7d","Type":"ContainerDied","Data":"916183b43ceb3e0ddda63423c2725edd4e1c6b8f21531521484baa210fcdcebf"} Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.434502 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.561567 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/121dd8c3-c175-4312-8d3c-4d3509908e7d-amphora-image\") pod \"121dd8c3-c175-4312-8d3c-4d3509908e7d\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.561669 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/121dd8c3-c175-4312-8d3c-4d3509908e7d-httpd-config\") pod \"121dd8c3-c175-4312-8d3c-4d3509908e7d\" (UID: \"121dd8c3-c175-4312-8d3c-4d3509908e7d\") " Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.595470 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/121dd8c3-c175-4312-8d3c-4d3509908e7d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "121dd8c3-c175-4312-8d3c-4d3509908e7d" (UID: "121dd8c3-c175-4312-8d3c-4d3509908e7d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.635086 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/121dd8c3-c175-4312-8d3c-4d3509908e7d-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "121dd8c3-c175-4312-8d3c-4d3509908e7d" (UID: "121dd8c3-c175-4312-8d3c-4d3509908e7d"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.680016 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/121dd8c3-c175-4312-8d3c-4d3509908e7d-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:12 crc kubenswrapper[4809]: I1205 12:48:12.680050 4809 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/121dd8c3-c175-4312-8d3c-4d3509908e7d-amphora-image\") on node \"crc\" DevicePath \"\"" Dec 05 12:48:13 crc kubenswrapper[4809]: I1205 12:48:13.086134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" event={"ID":"121dd8c3-c175-4312-8d3c-4d3509908e7d","Type":"ContainerDied","Data":"6b7d847552298ab23041f69db576c62b2d7c0e4034429712ba345b5084b3ddb6"} Dec 05 12:48:13 crc kubenswrapper[4809]: I1205 12:48:13.086439 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-kpf5q" Dec 05 12:48:13 crc kubenswrapper[4809]: I1205 12:48:13.086458 4809 scope.go:117] "RemoveContainer" containerID="916183b43ceb3e0ddda63423c2725edd4e1c6b8f21531521484baa210fcdcebf" Dec 05 12:48:13 crc kubenswrapper[4809]: I1205 12:48:13.124086 4809 scope.go:117] "RemoveContainer" containerID="79f8a08a450d1f37503e36ff68824b52b7e792ae9247ef02791daf8e912007b8" Dec 05 12:48:13 crc kubenswrapper[4809]: I1205 12:48:13.126978 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-kpf5q"] Dec 05 12:48:13 crc kubenswrapper[4809]: I1205 12:48:13.138582 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-kpf5q"] Dec 05 12:48:14 crc kubenswrapper[4809]: I1205 12:48:14.886330 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="121dd8c3-c175-4312-8d3c-4d3509908e7d" path="/var/lib/kubelet/pods/121dd8c3-c175-4312-8d3c-4d3509908e7d/volumes" Dec 05 12:48:16 crc kubenswrapper[4809]: I1205 12:48:16.873895 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.135230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"7e0ee3fefa7029bd441f660ece6169c58fb6143e423ed0039f9d6ad8d0c48340"} Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.180509 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-4gnqd"] Dec 05 12:48:17 crc kubenswrapper[4809]: E1205 12:48:17.181056 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1230fe8-b1b1-443e-adb6-37287dabebb6" containerName="init" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.181093 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1230fe8-b1b1-443e-adb6-37287dabebb6" containerName="init" Dec 05 12:48:17 crc kubenswrapper[4809]: E1205 12:48:17.181106 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerName="init" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.181115 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerName="init" Dec 05 12:48:17 crc kubenswrapper[4809]: E1205 12:48:17.181136 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1230fe8-b1b1-443e-adb6-37287dabebb6" containerName="octavia-db-sync" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.181145 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1230fe8-b1b1-443e-adb6-37287dabebb6" containerName="octavia-db-sync" Dec 05 12:48:17 crc kubenswrapper[4809]: E1205 12:48:17.181170 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerName="octavia-amphora-httpd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.181178 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerName="octavia-amphora-httpd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.181414 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="121dd8c3-c175-4312-8d3c-4d3509908e7d" containerName="octavia-amphora-httpd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.181448 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1230fe8-b1b1-443e-adb6-37287dabebb6" containerName="octavia-db-sync" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.182805 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.187462 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.194873 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-4gnqd"] Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.282720 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261-httpd-config\") pod \"octavia-image-upload-59f8cff499-4gnqd\" (UID: \"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261\") " pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.282911 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261-amphora-image\") pod \"octavia-image-upload-59f8cff499-4gnqd\" (UID: \"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261\") " pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.385185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261-httpd-config\") pod \"octavia-image-upload-59f8cff499-4gnqd\" (UID: \"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261\") " pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.385397 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261-amphora-image\") pod \"octavia-image-upload-59f8cff499-4gnqd\" (UID: \"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261\") " pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.386214 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261-amphora-image\") pod \"octavia-image-upload-59f8cff499-4gnqd\" (UID: \"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261\") " pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.393805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261-httpd-config\") pod \"octavia-image-upload-59f8cff499-4gnqd\" (UID: \"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261\") " pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:17 crc kubenswrapper[4809]: I1205 12:48:17.503545 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-4gnqd" Dec 05 12:48:18 crc kubenswrapper[4809]: W1205 12:48:18.023532 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bcbc9eb_9ab1_45d0_b7e3_b9ca40444261.slice/crio-27933e71bff18540873e6d491edfad7294026a6ba156a6304f20bfcddd43be1d WatchSource:0}: Error finding container 27933e71bff18540873e6d491edfad7294026a6ba156a6304f20bfcddd43be1d: Status 404 returned error can't find the container with id 27933e71bff18540873e6d491edfad7294026a6ba156a6304f20bfcddd43be1d Dec 05 12:48:18 crc kubenswrapper[4809]: I1205 12:48:18.025585 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-4gnqd"] Dec 05 12:48:18 crc kubenswrapper[4809]: I1205 12:48:18.149077 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-4gnqd" event={"ID":"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261","Type":"ContainerStarted","Data":"27933e71bff18540873e6d491edfad7294026a6ba156a6304f20bfcddd43be1d"} Dec 05 12:48:19 crc kubenswrapper[4809]: I1205 12:48:19.170444 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-4gnqd" event={"ID":"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261","Type":"ContainerStarted","Data":"053e0ac9e44d91d08e563bbaf1845e5b94f6852bf4b550be115effbb21c0208e"} Dec 05 12:48:20 crc kubenswrapper[4809]: I1205 12:48:20.183062 4809 generic.go:334] "Generic (PLEG): container finished" podID="6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261" containerID="053e0ac9e44d91d08e563bbaf1845e5b94f6852bf4b550be115effbb21c0208e" exitCode=0 Dec 05 12:48:20 crc kubenswrapper[4809]: I1205 12:48:20.183128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-4gnqd" event={"ID":"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261","Type":"ContainerDied","Data":"053e0ac9e44d91d08e563bbaf1845e5b94f6852bf4b550be115effbb21c0208e"} Dec 05 12:48:22 crc kubenswrapper[4809]: I1205 12:48:22.204843 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-4gnqd" event={"ID":"6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261","Type":"ContainerStarted","Data":"e877e317eec2aca4314c59e5a75db5b19f3e8ae684c599322f2f1307e6d9a782"} Dec 05 12:48:22 crc kubenswrapper[4809]: I1205 12:48:22.240884 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-4gnqd" podStartSLOduration=1.849538245 podStartE2EDuration="5.240856824s" podCreationTimestamp="2025-12-05 12:48:17 +0000 UTC" firstStartedPulling="2025-12-05 12:48:18.027182595 +0000 UTC m=+5993.418159153" lastFinishedPulling="2025-12-05 12:48:21.418501174 +0000 UTC m=+5996.809477732" observedRunningTime="2025-12-05 12:48:22.222599152 +0000 UTC m=+5997.613575730" watchObservedRunningTime="2025-12-05 12:48:22.240856824 +0000 UTC m=+5997.631833392" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.785135 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-4zc9q"] Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.787073 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.792428 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.792611 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.793045 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.814956 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-4zc9q"] Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.866021 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-config-data\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.866250 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/aff7629b-0b32-4e6a-999b-1a63d990acd9-hm-ports\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.866325 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-scripts\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.866548 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-combined-ca-bundle\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.866774 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-amphora-certs\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.867396 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aff7629b-0b32-4e6a-999b-1a63d990acd9-config-data-merged\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.969583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-combined-ca-bundle\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.970018 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-amphora-certs\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.970266 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aff7629b-0b32-4e6a-999b-1a63d990acd9-config-data-merged\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.970401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-config-data\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.970928 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aff7629b-0b32-4e6a-999b-1a63d990acd9-config-data-merged\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.971238 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/aff7629b-0b32-4e6a-999b-1a63d990acd9-hm-ports\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.971467 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-scripts\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.972231 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/aff7629b-0b32-4e6a-999b-1a63d990acd9-hm-ports\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.978495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-combined-ca-bundle\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.980188 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-scripts\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.985608 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-amphora-certs\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:29 crc kubenswrapper[4809]: I1205 12:48:29.993915 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aff7629b-0b32-4e6a-999b-1a63d990acd9-config-data\") pod \"octavia-healthmanager-4zc9q\" (UID: \"aff7629b-0b32-4e6a-999b-1a63d990acd9\") " pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:30 crc kubenswrapper[4809]: I1205 12:48:30.114995 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:30 crc kubenswrapper[4809]: I1205 12:48:30.842079 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-4zc9q"] Dec 05 12:48:30 crc kubenswrapper[4809]: W1205 12:48:30.846203 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaff7629b_0b32_4e6a_999b_1a63d990acd9.slice/crio-6609fe16d466edb97fa9ad06ba357914d44cbc81683795452f78f51da1af7351 WatchSource:0}: Error finding container 6609fe16d466edb97fa9ad06ba357914d44cbc81683795452f78f51da1af7351: Status 404 returned error can't find the container with id 6609fe16d466edb97fa9ad06ba357914d44cbc81683795452f78f51da1af7351 Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.296934 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-4zc9q" event={"ID":"aff7629b-0b32-4e6a-999b-1a63d990acd9","Type":"ContainerStarted","Data":"6609fe16d466edb97fa9ad06ba357914d44cbc81683795452f78f51da1af7351"} Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.373766 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-sf2cf"] Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.376404 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.379638 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.379775 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.385681 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-sf2cf"] Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.503100 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e42fdbf-f030-4f02-8912-e957f00db980-config-data-merged\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.503174 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-config-data\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.503215 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/3e42fdbf-f030-4f02-8912-e957f00db980-hm-ports\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.503553 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-scripts\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.503614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-amphora-certs\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.503868 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-combined-ca-bundle\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.605611 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e42fdbf-f030-4f02-8912-e957f00db980-config-data-merged\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.605739 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-config-data\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.605766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/3e42fdbf-f030-4f02-8912-e957f00db980-hm-ports\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.605883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-scripts\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.605911 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-amphora-certs\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.606023 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-combined-ca-bundle\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.606176 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e42fdbf-f030-4f02-8912-e957f00db980-config-data-merged\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.607414 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/3e42fdbf-f030-4f02-8912-e957f00db980-hm-ports\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.612821 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-scripts\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.616853 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-combined-ca-bundle\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.628469 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-amphora-certs\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.628942 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e42fdbf-f030-4f02-8912-e957f00db980-config-data\") pod \"octavia-housekeeping-sf2cf\" (UID: \"3e42fdbf-f030-4f02-8912-e957f00db980\") " pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:31 crc kubenswrapper[4809]: I1205 12:48:31.704125 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.232393 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-sf2cf"] Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.311230 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-4zc9q" event={"ID":"aff7629b-0b32-4e6a-999b-1a63d990acd9","Type":"ContainerStarted","Data":"0840e382d463022ab76f4619aec6e99a7d4eae820be29b34752a12bfeb970912"} Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.313128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-sf2cf" event={"ID":"3e42fdbf-f030-4f02-8912-e957f00db980","Type":"ContainerStarted","Data":"6fde7e1f07d05aa8d7757b1b3467f9ea64343611a70d974d0a84f3ffca5e13b4"} Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.357505 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-w28wb"] Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.386615 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.390347 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.400316 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.401590 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-w28wb"] Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.528301 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e414c222-8d57-46b8-bd74-36f1e9206dcf-hm-ports\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.528665 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-scripts\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.528887 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-combined-ca-bundle\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.528975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-config-data\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.529091 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-amphora-certs\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.529193 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e414c222-8d57-46b8-bd74-36f1e9206dcf-config-data-merged\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.630680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e414c222-8d57-46b8-bd74-36f1e9206dcf-hm-ports\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.631100 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-scripts\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.631253 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-combined-ca-bundle\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.631369 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-config-data\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.631482 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-amphora-certs\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.631582 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e414c222-8d57-46b8-bd74-36f1e9206dcf-config-data-merged\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.631899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e414c222-8d57-46b8-bd74-36f1e9206dcf-hm-ports\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.632607 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e414c222-8d57-46b8-bd74-36f1e9206dcf-config-data-merged\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.637818 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-scripts\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.638680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-config-data\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.639769 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-amphora-certs\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.644502 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e414c222-8d57-46b8-bd74-36f1e9206dcf-combined-ca-bundle\") pod \"octavia-worker-w28wb\" (UID: \"e414c222-8d57-46b8-bd74-36f1e9206dcf\") " pod="openstack/octavia-worker-w28wb" Dec 05 12:48:32 crc kubenswrapper[4809]: I1205 12:48:32.718910 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-w28wb" Dec 05 12:48:33 crc kubenswrapper[4809]: I1205 12:48:33.309418 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-w28wb"] Dec 05 12:48:33 crc kubenswrapper[4809]: W1205 12:48:33.486240 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode414c222_8d57_46b8_bd74_36f1e9206dcf.slice/crio-0771c7b025f0f641eaebea4f2ff037bc8030db5fd4d450711708a36209007bf4 WatchSource:0}: Error finding container 0771c7b025f0f641eaebea4f2ff037bc8030db5fd4d450711708a36209007bf4: Status 404 returned error can't find the container with id 0771c7b025f0f641eaebea4f2ff037bc8030db5fd4d450711708a36209007bf4 Dec 05 12:48:34 crc kubenswrapper[4809]: I1205 12:48:34.337941 4809 generic.go:334] "Generic (PLEG): container finished" podID="aff7629b-0b32-4e6a-999b-1a63d990acd9" containerID="0840e382d463022ab76f4619aec6e99a7d4eae820be29b34752a12bfeb970912" exitCode=0 Dec 05 12:48:34 crc kubenswrapper[4809]: I1205 12:48:34.337986 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-4zc9q" event={"ID":"aff7629b-0b32-4e6a-999b-1a63d990acd9","Type":"ContainerDied","Data":"0840e382d463022ab76f4619aec6e99a7d4eae820be29b34752a12bfeb970912"} Dec 05 12:48:34 crc kubenswrapper[4809]: I1205 12:48:34.339956 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w28wb" event={"ID":"e414c222-8d57-46b8-bd74-36f1e9206dcf","Type":"ContainerStarted","Data":"0771c7b025f0f641eaebea4f2ff037bc8030db5fd4d450711708a36209007bf4"} Dec 05 12:48:34 crc kubenswrapper[4809]: I1205 12:48:34.342054 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-sf2cf" event={"ID":"3e42fdbf-f030-4f02-8912-e957f00db980","Type":"ContainerStarted","Data":"b8d5681d0fedaa41969b20769edb56ecac901ee68d60211bbb517d1640470acc"} Dec 05 12:48:35 crc kubenswrapper[4809]: I1205 12:48:35.358103 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-4zc9q" event={"ID":"aff7629b-0b32-4e6a-999b-1a63d990acd9","Type":"ContainerStarted","Data":"b455111ba2682d4d863e0a4c1549367ada459bc24c06431bb24cfcde9b0ae651"} Dec 05 12:48:35 crc kubenswrapper[4809]: I1205 12:48:35.359416 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:35 crc kubenswrapper[4809]: I1205 12:48:35.361932 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w28wb" event={"ID":"e414c222-8d57-46b8-bd74-36f1e9206dcf","Type":"ContainerStarted","Data":"ebe6134b266e43d44ca8fbb1f19ec30522427e848e3ba16bb42aecea792d8805"} Dec 05 12:48:35 crc kubenswrapper[4809]: I1205 12:48:35.365624 4809 generic.go:334] "Generic (PLEG): container finished" podID="3e42fdbf-f030-4f02-8912-e957f00db980" containerID="b8d5681d0fedaa41969b20769edb56ecac901ee68d60211bbb517d1640470acc" exitCode=0 Dec 05 12:48:35 crc kubenswrapper[4809]: I1205 12:48:35.365674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-sf2cf" event={"ID":"3e42fdbf-f030-4f02-8912-e957f00db980","Type":"ContainerDied","Data":"b8d5681d0fedaa41969b20769edb56ecac901ee68d60211bbb517d1640470acc"} Dec 05 12:48:35 crc kubenswrapper[4809]: I1205 12:48:35.401162 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-4zc9q" podStartSLOduration=6.401140403 podStartE2EDuration="6.401140403s" podCreationTimestamp="2025-12-05 12:48:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:48:35.393710073 +0000 UTC m=+6010.784686651" watchObservedRunningTime="2025-12-05 12:48:35.401140403 +0000 UTC m=+6010.792116961" Dec 05 12:48:36 crc kubenswrapper[4809]: I1205 12:48:36.380605 4809 generic.go:334] "Generic (PLEG): container finished" podID="e414c222-8d57-46b8-bd74-36f1e9206dcf" containerID="ebe6134b266e43d44ca8fbb1f19ec30522427e848e3ba16bb42aecea792d8805" exitCode=0 Dec 05 12:48:36 crc kubenswrapper[4809]: I1205 12:48:36.380687 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w28wb" event={"ID":"e414c222-8d57-46b8-bd74-36f1e9206dcf","Type":"ContainerDied","Data":"ebe6134b266e43d44ca8fbb1f19ec30522427e848e3ba16bb42aecea792d8805"} Dec 05 12:48:36 crc kubenswrapper[4809]: I1205 12:48:36.385796 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-sf2cf" event={"ID":"3e42fdbf-f030-4f02-8912-e957f00db980","Type":"ContainerStarted","Data":"d57f02647376b250b3588161c95fb50d63ee0e209ca30c8715c81787ef8b40f4"} Dec 05 12:48:36 crc kubenswrapper[4809]: I1205 12:48:36.386570 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:36 crc kubenswrapper[4809]: I1205 12:48:36.448086 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-sf2cf" podStartSLOduration=4.145496824 podStartE2EDuration="5.448064026s" podCreationTimestamp="2025-12-05 12:48:31 +0000 UTC" firstStartedPulling="2025-12-05 12:48:32.241827397 +0000 UTC m=+6007.632803955" lastFinishedPulling="2025-12-05 12:48:33.544394589 +0000 UTC m=+6008.935371157" observedRunningTime="2025-12-05 12:48:36.42964798 +0000 UTC m=+6011.820624578" watchObservedRunningTime="2025-12-05 12:48:36.448064026 +0000 UTC m=+6011.839040584" Dec 05 12:48:37 crc kubenswrapper[4809]: I1205 12:48:37.397885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w28wb" event={"ID":"e414c222-8d57-46b8-bd74-36f1e9206dcf","Type":"ContainerStarted","Data":"fa3a6d25127a43fd06d504009718221015064922d0f08d89d905c5a686633413"} Dec 05 12:48:38 crc kubenswrapper[4809]: I1205 12:48:38.407561 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-w28wb" Dec 05 12:48:45 crc kubenswrapper[4809]: I1205 12:48:45.150021 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-4zc9q" Dec 05 12:48:45 crc kubenswrapper[4809]: I1205 12:48:45.169597 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-w28wb" podStartSLOduration=11.971581211 podStartE2EDuration="13.169574478s" podCreationTimestamp="2025-12-05 12:48:32 +0000 UTC" firstStartedPulling="2025-12-05 12:48:33.497098146 +0000 UTC m=+6008.888074704" lastFinishedPulling="2025-12-05 12:48:34.695091413 +0000 UTC m=+6010.086067971" observedRunningTime="2025-12-05 12:48:37.424750288 +0000 UTC m=+6012.815726866" watchObservedRunningTime="2025-12-05 12:48:45.169574478 +0000 UTC m=+6020.560551036" Dec 05 12:48:46 crc kubenswrapper[4809]: I1205 12:48:46.735161 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-sf2cf" Dec 05 12:48:47 crc kubenswrapper[4809]: I1205 12:48:47.749976 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-w28wb" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.304623 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ng7cz"] Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.325377 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.338433 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ng7cz"] Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.353800 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-catalog-content\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.353928 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-utilities\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.353975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqn8d\" (UniqueName: \"kubernetes.io/projected/6ac3692f-ad52-40cd-a177-3399f5dbb020-kube-api-access-nqn8d\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.456052 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-utilities\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.456352 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqn8d\" (UniqueName: \"kubernetes.io/projected/6ac3692f-ad52-40cd-a177-3399f5dbb020-kube-api-access-nqn8d\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.456559 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-catalog-content\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.456653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-utilities\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.456989 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-catalog-content\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.481816 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqn8d\" (UniqueName: \"kubernetes.io/projected/6ac3692f-ad52-40cd-a177-3399f5dbb020-kube-api-access-nqn8d\") pod \"certified-operators-ng7cz\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:02 crc kubenswrapper[4809]: I1205 12:49:02.645295 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:03 crc kubenswrapper[4809]: I1205 12:49:03.262007 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ng7cz"] Dec 05 12:49:03 crc kubenswrapper[4809]: I1205 12:49:03.711167 4809 generic.go:334] "Generic (PLEG): container finished" podID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerID="70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834" exitCode=0 Dec 05 12:49:03 crc kubenswrapper[4809]: I1205 12:49:03.711296 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ng7cz" event={"ID":"6ac3692f-ad52-40cd-a177-3399f5dbb020","Type":"ContainerDied","Data":"70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834"} Dec 05 12:49:03 crc kubenswrapper[4809]: I1205 12:49:03.712207 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ng7cz" event={"ID":"6ac3692f-ad52-40cd-a177-3399f5dbb020","Type":"ContainerStarted","Data":"d96aa0b47e5199b24c35d747f368f52df2f7f7bbb6b60a31178fe8da94648658"} Dec 05 12:49:04 crc kubenswrapper[4809]: I1205 12:49:04.724730 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ng7cz" event={"ID":"6ac3692f-ad52-40cd-a177-3399f5dbb020","Type":"ContainerStarted","Data":"6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748"} Dec 05 12:49:05 crc kubenswrapper[4809]: I1205 12:49:05.739610 4809 generic.go:334] "Generic (PLEG): container finished" podID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerID="6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748" exitCode=0 Dec 05 12:49:05 crc kubenswrapper[4809]: I1205 12:49:05.739674 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ng7cz" event={"ID":"6ac3692f-ad52-40cd-a177-3399f5dbb020","Type":"ContainerDied","Data":"6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748"} Dec 05 12:49:06 crc kubenswrapper[4809]: I1205 12:49:06.751113 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ng7cz" event={"ID":"6ac3692f-ad52-40cd-a177-3399f5dbb020","Type":"ContainerStarted","Data":"f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27"} Dec 05 12:49:06 crc kubenswrapper[4809]: I1205 12:49:06.772280 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ng7cz" podStartSLOduration=2.332885638 podStartE2EDuration="4.772261421s" podCreationTimestamp="2025-12-05 12:49:02 +0000 UTC" firstStartedPulling="2025-12-05 12:49:03.714289942 +0000 UTC m=+6039.105266500" lastFinishedPulling="2025-12-05 12:49:06.153665725 +0000 UTC m=+6041.544642283" observedRunningTime="2025-12-05 12:49:06.768834559 +0000 UTC m=+6042.159811137" watchObservedRunningTime="2025-12-05 12:49:06.772261421 +0000 UTC m=+6042.163237979" Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.048486 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rl8qz"] Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.059985 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rl8qz"] Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.070359 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-cba5-account-create-update-kmm8r"] Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.080425 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-cba5-account-create-update-kmm8r"] Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.646278 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.646640 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.699537 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.863957 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.884003 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ec410a7-ff97-4658-a11f-dcf2de86d676" path="/var/lib/kubelet/pods/7ec410a7-ff97-4658-a11f-dcf2de86d676/volumes" Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.884978 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea306765-6369-43b9-85ba-cb8046c8c489" path="/var/lib/kubelet/pods/ea306765-6369-43b9-85ba-cb8046c8c489/volumes" Dec 05 12:49:12 crc kubenswrapper[4809]: I1205 12:49:12.936985 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ng7cz"] Dec 05 12:49:14 crc kubenswrapper[4809]: I1205 12:49:14.837804 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ng7cz" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="registry-server" containerID="cri-o://f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27" gracePeriod=2 Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.309656 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.417166 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-utilities\") pod \"6ac3692f-ad52-40cd-a177-3399f5dbb020\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.417218 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqn8d\" (UniqueName: \"kubernetes.io/projected/6ac3692f-ad52-40cd-a177-3399f5dbb020-kube-api-access-nqn8d\") pod \"6ac3692f-ad52-40cd-a177-3399f5dbb020\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.417300 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-catalog-content\") pod \"6ac3692f-ad52-40cd-a177-3399f5dbb020\" (UID: \"6ac3692f-ad52-40cd-a177-3399f5dbb020\") " Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.417971 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-utilities" (OuterVolumeSpecName: "utilities") pod "6ac3692f-ad52-40cd-a177-3399f5dbb020" (UID: "6ac3692f-ad52-40cd-a177-3399f5dbb020"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.436170 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac3692f-ad52-40cd-a177-3399f5dbb020-kube-api-access-nqn8d" (OuterVolumeSpecName: "kube-api-access-nqn8d") pod "6ac3692f-ad52-40cd-a177-3399f5dbb020" (UID: "6ac3692f-ad52-40cd-a177-3399f5dbb020"). InnerVolumeSpecName "kube-api-access-nqn8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.468102 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ac3692f-ad52-40cd-a177-3399f5dbb020" (UID: "6ac3692f-ad52-40cd-a177-3399f5dbb020"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.519536 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.519776 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqn8d\" (UniqueName: \"kubernetes.io/projected/6ac3692f-ad52-40cd-a177-3399f5dbb020-kube-api-access-nqn8d\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.519791 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac3692f-ad52-40cd-a177-3399f5dbb020-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.849790 4809 generic.go:334] "Generic (PLEG): container finished" podID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerID="f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27" exitCode=0 Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.849836 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ng7cz" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.849832 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ng7cz" event={"ID":"6ac3692f-ad52-40cd-a177-3399f5dbb020","Type":"ContainerDied","Data":"f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27"} Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.849880 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ng7cz" event={"ID":"6ac3692f-ad52-40cd-a177-3399f5dbb020","Type":"ContainerDied","Data":"d96aa0b47e5199b24c35d747f368f52df2f7f7bbb6b60a31178fe8da94648658"} Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.849903 4809 scope.go:117] "RemoveContainer" containerID="f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.881187 4809 scope.go:117] "RemoveContainer" containerID="6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.890308 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ng7cz"] Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.898671 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ng7cz"] Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.905934 4809 scope.go:117] "RemoveContainer" containerID="70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.971689 4809 scope.go:117] "RemoveContainer" containerID="f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27" Dec 05 12:49:15 crc kubenswrapper[4809]: E1205 12:49:15.972167 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27\": container with ID starting with f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27 not found: ID does not exist" containerID="f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.972207 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27"} err="failed to get container status \"f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27\": rpc error: code = NotFound desc = could not find container \"f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27\": container with ID starting with f482d9a86ce4b63d25b836227384cebefe042658e96e034970cce8ca46defd27 not found: ID does not exist" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.972231 4809 scope.go:117] "RemoveContainer" containerID="6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748" Dec 05 12:49:15 crc kubenswrapper[4809]: E1205 12:49:15.972589 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748\": container with ID starting with 6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748 not found: ID does not exist" containerID="6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.972613 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748"} err="failed to get container status \"6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748\": rpc error: code = NotFound desc = could not find container \"6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748\": container with ID starting with 6fa5133c55a2abea9213327e1eed803ec1d63bb85e3f6f8c8bf3ad92062f9748 not found: ID does not exist" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.972628 4809 scope.go:117] "RemoveContainer" containerID="70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834" Dec 05 12:49:15 crc kubenswrapper[4809]: E1205 12:49:15.972968 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834\": container with ID starting with 70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834 not found: ID does not exist" containerID="70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834" Dec 05 12:49:15 crc kubenswrapper[4809]: I1205 12:49:15.972996 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834"} err="failed to get container status \"70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834\": rpc error: code = NotFound desc = could not find container \"70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834\": container with ID starting with 70849ca4fc577d3663952670e74f05d7aad151e46def588c11cd1870cb453834 not found: ID does not exist" Dec 05 12:49:16 crc kubenswrapper[4809]: I1205 12:49:16.887007 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" path="/var/lib/kubelet/pods/6ac3692f-ad52-40cd-a177-3399f5dbb020/volumes" Dec 05 12:49:19 crc kubenswrapper[4809]: I1205 12:49:19.044057 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-hz87m"] Dec 05 12:49:19 crc kubenswrapper[4809]: I1205 12:49:19.052454 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-hz87m"] Dec 05 12:49:20 crc kubenswrapper[4809]: I1205 12:49:20.885279 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d65186f3-26d3-4c29-8bc8-41d0b57891a1" path="/var/lib/kubelet/pods/d65186f3-26d3-4c29-8bc8-41d0b57891a1/volumes" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.618338 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-778c96f7-5qmgr"] Dec 05 12:49:34 crc kubenswrapper[4809]: E1205 12:49:34.620567 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="extract-utilities" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.620737 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="extract-utilities" Dec 05 12:49:34 crc kubenswrapper[4809]: E1205 12:49:34.620861 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="registry-server" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.620941 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="registry-server" Dec 05 12:49:34 crc kubenswrapper[4809]: E1205 12:49:34.621023 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="extract-content" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.621105 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="extract-content" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.621417 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac3692f-ad52-40cd-a177-3399f5dbb020" containerName="registry-server" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.622818 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.629598 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.629897 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-gmjqv" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.630124 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.630259 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.636363 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-778c96f7-5qmgr"] Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.677795 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.678088 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-log" containerID="cri-o://825928806dff97aafe8db5614b26d9c6e3400f3755111b41a90eaf2c5c096d56" gracePeriod=30 Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.678725 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-httpd" containerID="cri-o://7e9e0d1c4054e24fb71232d57d5d2ed077ca1fcafa3b53fcda48f9701a1b0aa8" gracePeriod=30 Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.748307 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.748760 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-log" containerID="cri-o://f9a990752c48371d529e3c5e427ac07606f04f0669e1a6d399824b2077f1a7a8" gracePeriod=30 Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.749204 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-httpd" containerID="cri-o://295010b98790218c4d1fe7af3615987df15f65937024f30689173d2a697cb687" gracePeriod=30 Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.765353 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79969cdbc9-jcmsj"] Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.767450 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820151 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8556f0ed-27ba-4369-9a2b-325753529f06-logs\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820243 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-config-data\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820283 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd3251ad-f094-4ab9-93b0-59c26aff5a40-horizon-secret-key\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8556f0ed-27ba-4369-9a2b-325753529f06-horizon-secret-key\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820391 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54pkq\" (UniqueName: \"kubernetes.io/projected/dd3251ad-f094-4ab9-93b0-59c26aff5a40-kube-api-access-54pkq\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820481 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd3251ad-f094-4ab9-93b0-59c26aff5a40-logs\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820557 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-scripts\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820650 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-config-data\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820756 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-scripts\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.820892 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9fws\" (UniqueName: \"kubernetes.io/projected/8556f0ed-27ba-4369-9a2b-325753529f06-kube-api-access-d9fws\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.838085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79969cdbc9-jcmsj"] Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.922980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-config-data\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-scripts\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923156 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9fws\" (UniqueName: \"kubernetes.io/projected/8556f0ed-27ba-4369-9a2b-325753529f06-kube-api-access-d9fws\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8556f0ed-27ba-4369-9a2b-325753529f06-logs\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923262 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-config-data\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923295 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd3251ad-f094-4ab9-93b0-59c26aff5a40-horizon-secret-key\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923336 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8556f0ed-27ba-4369-9a2b-325753529f06-horizon-secret-key\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923371 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54pkq\" (UniqueName: \"kubernetes.io/projected/dd3251ad-f094-4ab9-93b0-59c26aff5a40-kube-api-access-54pkq\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923420 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd3251ad-f094-4ab9-93b0-59c26aff5a40-logs\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-scripts\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.923808 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8556f0ed-27ba-4369-9a2b-325753529f06-logs\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.924426 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd3251ad-f094-4ab9-93b0-59c26aff5a40-logs\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.925424 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-scripts\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.925616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-config-data\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.925679 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-scripts\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.926897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-config-data\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.930535 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8556f0ed-27ba-4369-9a2b-325753529f06-horizon-secret-key\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.930581 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd3251ad-f094-4ab9-93b0-59c26aff5a40-horizon-secret-key\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.941979 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54pkq\" (UniqueName: \"kubernetes.io/projected/dd3251ad-f094-4ab9-93b0-59c26aff5a40-kube-api-access-54pkq\") pod \"horizon-778c96f7-5qmgr\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.942107 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9fws\" (UniqueName: \"kubernetes.io/projected/8556f0ed-27ba-4369-9a2b-325753529f06-kube-api-access-d9fws\") pod \"horizon-79969cdbc9-jcmsj\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:34 crc kubenswrapper[4809]: I1205 12:49:34.962316 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.017131 4809 generic.go:334] "Generic (PLEG): container finished" podID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerID="825928806dff97aafe8db5614b26d9c6e3400f3755111b41a90eaf2c5c096d56" exitCode=143 Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.017291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b435d598-7879-4eeb-97fa-91d8df2ec977","Type":"ContainerDied","Data":"825928806dff97aafe8db5614b26d9c6e3400f3755111b41a90eaf2c5c096d56"} Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.023792 4809 generic.go:334] "Generic (PLEG): container finished" podID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerID="f9a990752c48371d529e3c5e427ac07606f04f0669e1a6d399824b2077f1a7a8" exitCode=143 Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.023839 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e","Type":"ContainerDied","Data":"f9a990752c48371d529e3c5e427ac07606f04f0669e1a6d399824b2077f1a7a8"} Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.202084 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.428129 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-778c96f7-5qmgr"] Dec 05 12:49:35 crc kubenswrapper[4809]: W1205 12:49:35.437032 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd3251ad_f094_4ab9_93b0_59c26aff5a40.slice/crio-014411af6d785d95a1f0dc8e3bc4de6e4ce967dad903634dfe0d9b456ddd446d WatchSource:0}: Error finding container 014411af6d785d95a1f0dc8e3bc4de6e4ce967dad903634dfe0d9b456ddd446d: Status 404 returned error can't find the container with id 014411af6d785d95a1f0dc8e3bc4de6e4ce967dad903634dfe0d9b456ddd446d Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.457439 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-778c96f7-5qmgr"] Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.495161 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-855b9549c7-7gz4j"] Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.497288 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.505924 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-855b9549c7-7gz4j"] Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.536586 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-scripts\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.536785 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2373ada0-4330-4520-a4a8-975f8be6a4f3-horizon-secret-key\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.536836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373ada0-4330-4520-a4a8-975f8be6a4f3-logs\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.536871 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps2kn\" (UniqueName: \"kubernetes.io/projected/2373ada0-4330-4520-a4a8-975f8be6a4f3-kube-api-access-ps2kn\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.536922 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-config-data\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.639011 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2373ada0-4330-4520-a4a8-975f8be6a4f3-horizon-secret-key\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.639081 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373ada0-4330-4520-a4a8-975f8be6a4f3-logs\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.639117 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps2kn\" (UniqueName: \"kubernetes.io/projected/2373ada0-4330-4520-a4a8-975f8be6a4f3-kube-api-access-ps2kn\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.639162 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-config-data\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.639207 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-scripts\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.640011 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-scripts\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.640691 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373ada0-4330-4520-a4a8-975f8be6a4f3-logs\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.640958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-config-data\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.646403 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2373ada0-4330-4520-a4a8-975f8be6a4f3-horizon-secret-key\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.659446 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps2kn\" (UniqueName: \"kubernetes.io/projected/2373ada0-4330-4520-a4a8-975f8be6a4f3-kube-api-access-ps2kn\") pod \"horizon-855b9549c7-7gz4j\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.669411 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79969cdbc9-jcmsj"] Dec 05 12:49:35 crc kubenswrapper[4809]: W1205 12:49:35.670164 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8556f0ed_27ba_4369_9a2b_325753529f06.slice/crio-bc12d0c1f1c0d29fd10297c7d3f3e89b51e667a1a34851c87da413bf0ab9f4b9 WatchSource:0}: Error finding container bc12d0c1f1c0d29fd10297c7d3f3e89b51e667a1a34851c87da413bf0ab9f4b9: Status 404 returned error can't find the container with id bc12d0c1f1c0d29fd10297c7d3f3e89b51e667a1a34851c87da413bf0ab9f4b9 Dec 05 12:49:35 crc kubenswrapper[4809]: I1205 12:49:35.818075 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:36 crc kubenswrapper[4809]: I1205 12:49:36.036159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79969cdbc9-jcmsj" event={"ID":"8556f0ed-27ba-4369-9a2b-325753529f06","Type":"ContainerStarted","Data":"bc12d0c1f1c0d29fd10297c7d3f3e89b51e667a1a34851c87da413bf0ab9f4b9"} Dec 05 12:49:36 crc kubenswrapper[4809]: I1205 12:49:36.037756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-778c96f7-5qmgr" event={"ID":"dd3251ad-f094-4ab9-93b0-59c26aff5a40","Type":"ContainerStarted","Data":"014411af6d785d95a1f0dc8e3bc4de6e4ce967dad903634dfe0d9b456ddd446d"} Dec 05 12:49:36 crc kubenswrapper[4809]: I1205 12:49:36.295098 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-855b9549c7-7gz4j"] Dec 05 12:49:37 crc kubenswrapper[4809]: I1205 12:49:37.048492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855b9549c7-7gz4j" event={"ID":"2373ada0-4330-4520-a4a8-975f8be6a4f3","Type":"ContainerStarted","Data":"6b91b3fe82dcfb2989f193135e2f1f8a9bc66aded2b8b4f3e05e71d92586c8e7"} Dec 05 12:49:38 crc kubenswrapper[4809]: I1205 12:49:38.059343 4809 generic.go:334] "Generic (PLEG): container finished" podID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerID="295010b98790218c4d1fe7af3615987df15f65937024f30689173d2a697cb687" exitCode=0 Dec 05 12:49:38 crc kubenswrapper[4809]: I1205 12:49:38.059611 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e","Type":"ContainerDied","Data":"295010b98790218c4d1fe7af3615987df15f65937024f30689173d2a697cb687"} Dec 05 12:49:38 crc kubenswrapper[4809]: I1205 12:49:38.061490 4809 generic.go:334] "Generic (PLEG): container finished" podID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerID="7e9e0d1c4054e24fb71232d57d5d2ed077ca1fcafa3b53fcda48f9701a1b0aa8" exitCode=0 Dec 05 12:49:38 crc kubenswrapper[4809]: I1205 12:49:38.061527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b435d598-7879-4eeb-97fa-91d8df2ec977","Type":"ContainerDied","Data":"7e9e0d1c4054e24fb71232d57d5d2ed077ca1fcafa3b53fcda48f9701a1b0aa8"} Dec 05 12:49:38 crc kubenswrapper[4809]: I1205 12:49:38.740703 4809 scope.go:117] "RemoveContainer" containerID="2e74766c5d5b2b280f7d796bfbe66eb20ff3d9a4c1350f688cebedd4f4956608" Dec 05 12:49:41 crc kubenswrapper[4809]: I1205 12:49:41.774788 4809 scope.go:117] "RemoveContainer" containerID="4ccb7977ea8f4338e5eb061e90aad234bc5a6744cd141eb8c3be3b6c57df63bd" Dec 05 12:49:41 crc kubenswrapper[4809]: I1205 12:49:41.880479 4809 scope.go:117] "RemoveContainer" containerID="83b504189f41cf2291b81c492f373c8072043fb0be2b9079538fe4f8c64ab7a7" Dec 05 12:49:41 crc kubenswrapper[4809]: I1205 12:49:41.934515 4809 scope.go:117] "RemoveContainer" containerID="d1780c19b16939eeaaa8bc199ed0de29a951741ba1343c869efa054a59398bb7" Dec 05 12:49:41 crc kubenswrapper[4809]: I1205 12:49:41.981764 4809 scope.go:117] "RemoveContainer" containerID="e46ecbca15b989fbf4b4d0b601955eb37da770af3bdb56b624aa574edd1b6d4e" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.234821 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.340219 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.376264 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-httpd-run\") pod \"b435d598-7879-4eeb-97fa-91d8df2ec977\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.376370 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-combined-ca-bundle\") pod \"b435d598-7879-4eeb-97fa-91d8df2ec977\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.376420 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-config-data\") pod \"b435d598-7879-4eeb-97fa-91d8df2ec977\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.376449 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t47hv\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-kube-api-access-t47hv\") pod \"b435d598-7879-4eeb-97fa-91d8df2ec977\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.376487 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-logs\") pod \"b435d598-7879-4eeb-97fa-91d8df2ec977\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.376515 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-ceph\") pod \"b435d598-7879-4eeb-97fa-91d8df2ec977\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.376537 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-scripts\") pod \"b435d598-7879-4eeb-97fa-91d8df2ec977\" (UID: \"b435d598-7879-4eeb-97fa-91d8df2ec977\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.382424 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-kube-api-access-t47hv" (OuterVolumeSpecName: "kube-api-access-t47hv") pod "b435d598-7879-4eeb-97fa-91d8df2ec977" (UID: "b435d598-7879-4eeb-97fa-91d8df2ec977"). InnerVolumeSpecName "kube-api-access-t47hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.382938 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b435d598-7879-4eeb-97fa-91d8df2ec977" (UID: "b435d598-7879-4eeb-97fa-91d8df2ec977"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.385228 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-scripts" (OuterVolumeSpecName: "scripts") pod "b435d598-7879-4eeb-97fa-91d8df2ec977" (UID: "b435d598-7879-4eeb-97fa-91d8df2ec977"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.385402 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-logs" (OuterVolumeSpecName: "logs") pod "b435d598-7879-4eeb-97fa-91d8df2ec977" (UID: "b435d598-7879-4eeb-97fa-91d8df2ec977"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.391043 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-ceph" (OuterVolumeSpecName: "ceph") pod "b435d598-7879-4eeb-97fa-91d8df2ec977" (UID: "b435d598-7879-4eeb-97fa-91d8df2ec977"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.431305 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b435d598-7879-4eeb-97fa-91d8df2ec977" (UID: "b435d598-7879-4eeb-97fa-91d8df2ec977"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.473177 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-config-data" (OuterVolumeSpecName: "config-data") pod "b435d598-7879-4eeb-97fa-91d8df2ec977" (UID: "b435d598-7879-4eeb-97fa-91d8df2ec977"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.477671 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-httpd-run\") pod \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.477951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-config-data\") pod \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.478052 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-combined-ca-bundle\") pod \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.478170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-scripts\") pod \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.478265 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-ceph\") pod \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.478362 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s46kl\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-kube-api-access-s46kl\") pod \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.478495 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-logs\") pod \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\" (UID: \"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e\") " Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479403 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479503 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479594 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479700 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t47hv\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-kube-api-access-t47hv\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479785 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b435d598-7879-4eeb-97fa-91d8df2ec977-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479860 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b435d598-7879-4eeb-97fa-91d8df2ec977-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479957 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b435d598-7879-4eeb-97fa-91d8df2ec977-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.479779 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" (UID: "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.480438 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-logs" (OuterVolumeSpecName: "logs") pod "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" (UID: "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.487491 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-kube-api-access-s46kl" (OuterVolumeSpecName: "kube-api-access-s46kl") pod "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" (UID: "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e"). InnerVolumeSpecName "kube-api-access-s46kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.487585 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-ceph" (OuterVolumeSpecName: "ceph") pod "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" (UID: "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.487650 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-scripts" (OuterVolumeSpecName: "scripts") pod "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" (UID: "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.512712 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" (UID: "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.542309 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-config-data" (OuterVolumeSpecName: "config-data") pod "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" (UID: "fbf81b4e-4c9d-4ee9-8149-fc17884aa78e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.581973 4809 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.582001 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.582011 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.582020 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.582028 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.582036 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s46kl\" (UniqueName: \"kubernetes.io/projected/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-kube-api-access-s46kl\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:42 crc kubenswrapper[4809]: I1205 12:49:42.582045 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.136209 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79969cdbc9-jcmsj" event={"ID":"8556f0ed-27ba-4369-9a2b-325753529f06","Type":"ContainerStarted","Data":"a1aca2ec9e8244d58893e3df45a37d7a1e87013edcaf1c0e09d5699b5157b059"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.136514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79969cdbc9-jcmsj" event={"ID":"8556f0ed-27ba-4369-9a2b-325753529f06","Type":"ContainerStarted","Data":"90ee1c66a6078532b65d785c6b746344291f4193843178889388821d0c5653e0"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.151116 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-778c96f7-5qmgr" event={"ID":"dd3251ad-f094-4ab9-93b0-59c26aff5a40","Type":"ContainerStarted","Data":"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.151330 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-778c96f7-5qmgr" event={"ID":"dd3251ad-f094-4ab9-93b0-59c26aff5a40","Type":"ContainerStarted","Data":"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.151559 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-778c96f7-5qmgr" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon-log" containerID="cri-o://ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86" gracePeriod=30 Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.151802 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-778c96f7-5qmgr" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon" containerID="cri-o://6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb" gracePeriod=30 Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.170706 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b435d598-7879-4eeb-97fa-91d8df2ec977","Type":"ContainerDied","Data":"5b74067b930ef3e0a28930e6256c29e661bd4876b580eea7a64c79bf6e623cd6"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.170753 4809 scope.go:117] "RemoveContainer" containerID="7e9e0d1c4054e24fb71232d57d5d2ed077ca1fcafa3b53fcda48f9701a1b0aa8" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.170881 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.187006 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79969cdbc9-jcmsj" podStartSLOduration=2.913470473 podStartE2EDuration="9.186976511s" podCreationTimestamp="2025-12-05 12:49:34 +0000 UTC" firstStartedPulling="2025-12-05 12:49:35.6736877 +0000 UTC m=+6071.064664258" lastFinishedPulling="2025-12-05 12:49:41.947193738 +0000 UTC m=+6077.338170296" observedRunningTime="2025-12-05 12:49:43.164169897 +0000 UTC m=+6078.555146465" watchObservedRunningTime="2025-12-05 12:49:43.186976511 +0000 UTC m=+6078.577953069" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.219030 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855b9549c7-7gz4j" event={"ID":"2373ada0-4330-4520-a4a8-975f8be6a4f3","Type":"ContainerStarted","Data":"69be850a99b8866ddc464c1264d39c7f17f9f7f430a0d7e677d274f2b5baf3c9"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.219089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855b9549c7-7gz4j" event={"ID":"2373ada0-4330-4520-a4a8-975f8be6a4f3","Type":"ContainerStarted","Data":"410fa6e02707d6e1a8e040478cafffa70737c869e9fe716ccbb6a2b162601460"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.229051 4809 scope.go:117] "RemoveContainer" containerID="825928806dff97aafe8db5614b26d9c6e3400f3755111b41a90eaf2c5c096d56" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.243890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fbf81b4e-4c9d-4ee9-8149-fc17884aa78e","Type":"ContainerDied","Data":"9d03496bc2ff9d0d6d8977764b2367ac63eff8fef327619da24bb1fd7b4db617"} Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.243998 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.245551 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-778c96f7-5qmgr" podStartSLOduration=2.766409565 podStartE2EDuration="9.245529906s" podCreationTimestamp="2025-12-05 12:49:34 +0000 UTC" firstStartedPulling="2025-12-05 12:49:35.438582293 +0000 UTC m=+6070.829558851" lastFinishedPulling="2025-12-05 12:49:41.917702634 +0000 UTC m=+6077.308679192" observedRunningTime="2025-12-05 12:49:43.211368007 +0000 UTC m=+6078.602344575" watchObservedRunningTime="2025-12-05 12:49:43.245529906 +0000 UTC m=+6078.636506464" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.249036 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.268605 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.275845 4809 scope.go:117] "RemoveContainer" containerID="295010b98790218c4d1fe7af3615987df15f65937024f30689173d2a697cb687" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.280182 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: E1205 12:49:43.280693 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-log" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.280719 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-log" Dec 05 12:49:43 crc kubenswrapper[4809]: E1205 12:49:43.280762 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-httpd" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.280771 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-httpd" Dec 05 12:49:43 crc kubenswrapper[4809]: E1205 12:49:43.280788 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-log" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.280797 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-log" Dec 05 12:49:43 crc kubenswrapper[4809]: E1205 12:49:43.280810 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-httpd" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.280818 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-httpd" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.281078 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-httpd" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.281102 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-log" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.281138 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" containerName="glance-httpd" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.281150 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" containerName="glance-log" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.281469 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-855b9549c7-7gz4j" podStartSLOduration=2.64918912 podStartE2EDuration="8.281451633s" podCreationTimestamp="2025-12-05 12:49:35 +0000 UTC" firstStartedPulling="2025-12-05 12:49:36.309273193 +0000 UTC m=+6071.700249751" lastFinishedPulling="2025-12-05 12:49:41.941535696 +0000 UTC m=+6077.332512264" observedRunningTime="2025-12-05 12:49:43.27170206 +0000 UTC m=+6078.662678618" watchObservedRunningTime="2025-12-05 12:49:43.281451633 +0000 UTC m=+6078.672428201" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.282379 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.288823 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.290088 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.291883 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-nhzj2" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.330673 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.350125 4809 scope.go:117] "RemoveContainer" containerID="f9a990752c48371d529e3c5e427ac07606f04f0669e1a6d399824b2077f1a7a8" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.363807 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.379424 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.400608 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/323c1d96-4a17-48b4-a300-57dd30a08cec-logs\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.400708 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xzgv\" (UniqueName: \"kubernetes.io/projected/323c1d96-4a17-48b4-a300-57dd30a08cec-kube-api-access-9xzgv\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.400746 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-config-data\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.400891 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/323c1d96-4a17-48b4-a300-57dd30a08cec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.400927 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.400989 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/323c1d96-4a17-48b4-a300-57dd30a08cec-ceph\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.401070 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-scripts\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.408265 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.410411 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.412402 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.418296 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.502890 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0efb9902-75fb-4f9b-8966-8b70f61cd523-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503003 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503069 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efb9902-75fb-4f9b-8966-8b70f61cd523-logs\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/323c1d96-4a17-48b4-a300-57dd30a08cec-logs\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503160 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0efb9902-75fb-4f9b-8966-8b70f61cd523-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/323c1d96-4a17-48b4-a300-57dd30a08cec-logs\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503177 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503780 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xzgv\" (UniqueName: \"kubernetes.io/projected/323c1d96-4a17-48b4-a300-57dd30a08cec-kube-api-access-9xzgv\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.503803 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-config-data\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.504099 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/323c1d96-4a17-48b4-a300-57dd30a08cec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.504378 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/323c1d96-4a17-48b4-a300-57dd30a08cec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.504134 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.504464 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/323c1d96-4a17-48b4-a300-57dd30a08cec-ceph\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.505054 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7pfq\" (UniqueName: \"kubernetes.io/projected/0efb9902-75fb-4f9b-8966-8b70f61cd523-kube-api-access-f7pfq\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.505080 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-scripts\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.505195 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.508887 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-scripts\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.509578 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/323c1d96-4a17-48b4-a300-57dd30a08cec-ceph\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.510146 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-config-data\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.527931 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/323c1d96-4a17-48b4-a300-57dd30a08cec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.530538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xzgv\" (UniqueName: \"kubernetes.io/projected/323c1d96-4a17-48b4-a300-57dd30a08cec-kube-api-access-9xzgv\") pod \"glance-default-external-api-0\" (UID: \"323c1d96-4a17-48b4-a300-57dd30a08cec\") " pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.607589 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0efb9902-75fb-4f9b-8966-8b70f61cd523-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.608043 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0efb9902-75fb-4f9b-8966-8b70f61cd523-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.608165 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.608309 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7pfq\" (UniqueName: \"kubernetes.io/projected/0efb9902-75fb-4f9b-8966-8b70f61cd523-kube-api-access-f7pfq\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.608338 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.609039 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0efb9902-75fb-4f9b-8966-8b70f61cd523-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.609088 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.609112 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efb9902-75fb-4f9b-8966-8b70f61cd523-logs\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.609387 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efb9902-75fb-4f9b-8966-8b70f61cd523-logs\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.612487 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.612832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.613731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0efb9902-75fb-4f9b-8966-8b70f61cd523-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.618532 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efb9902-75fb-4f9b-8966-8b70f61cd523-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.620649 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.627608 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7pfq\" (UniqueName: \"kubernetes.io/projected/0efb9902-75fb-4f9b-8966-8b70f61cd523-kube-api-access-f7pfq\") pod \"glance-default-internal-api-0\" (UID: \"0efb9902-75fb-4f9b-8966-8b70f61cd523\") " pod="openstack/glance-default-internal-api-0" Dec 05 12:49:43 crc kubenswrapper[4809]: I1205 12:49:43.736196 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:44 crc kubenswrapper[4809]: I1205 12:49:44.173825 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 12:49:44 crc kubenswrapper[4809]: W1205 12:49:44.180288 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod323c1d96_4a17_48b4_a300_57dd30a08cec.slice/crio-3f1c2d67eeeb2758a0df68c269ad99a3365f59ada9588a18f679e4bdfd3e24de WatchSource:0}: Error finding container 3f1c2d67eeeb2758a0df68c269ad99a3365f59ada9588a18f679e4bdfd3e24de: Status 404 returned error can't find the container with id 3f1c2d67eeeb2758a0df68c269ad99a3365f59ada9588a18f679e4bdfd3e24de Dec 05 12:49:44 crc kubenswrapper[4809]: I1205 12:49:44.287504 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"323c1d96-4a17-48b4-a300-57dd30a08cec","Type":"ContainerStarted","Data":"3f1c2d67eeeb2758a0df68c269ad99a3365f59ada9588a18f679e4bdfd3e24de"} Dec 05 12:49:44 crc kubenswrapper[4809]: I1205 12:49:44.365924 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 12:49:44 crc kubenswrapper[4809]: W1205 12:49:44.386614 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0efb9902_75fb_4f9b_8966_8b70f61cd523.slice/crio-805fdb3d02a645af3dc10dbb81c90cedb3090d166822c707fc5e65585e69e7b6 WatchSource:0}: Error finding container 805fdb3d02a645af3dc10dbb81c90cedb3090d166822c707fc5e65585e69e7b6: Status 404 returned error can't find the container with id 805fdb3d02a645af3dc10dbb81c90cedb3090d166822c707fc5e65585e69e7b6 Dec 05 12:49:44 crc kubenswrapper[4809]: I1205 12:49:44.907031 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b435d598-7879-4eeb-97fa-91d8df2ec977" path="/var/lib/kubelet/pods/b435d598-7879-4eeb-97fa-91d8df2ec977/volumes" Dec 05 12:49:44 crc kubenswrapper[4809]: I1205 12:49:44.908597 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbf81b4e-4c9d-4ee9-8149-fc17884aa78e" path="/var/lib/kubelet/pods/fbf81b4e-4c9d-4ee9-8149-fc17884aa78e/volumes" Dec 05 12:49:44 crc kubenswrapper[4809]: I1205 12:49:44.969069 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:49:45 crc kubenswrapper[4809]: I1205 12:49:45.203429 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:45 crc kubenswrapper[4809]: I1205 12:49:45.203774 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:49:45 crc kubenswrapper[4809]: I1205 12:49:45.301332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"323c1d96-4a17-48b4-a300-57dd30a08cec","Type":"ContainerStarted","Data":"924b57ed620a2aa5ae2a27747021dd4f429777512c0390fc35c089ac2d3f2fa0"} Dec 05 12:49:45 crc kubenswrapper[4809]: I1205 12:49:45.304079 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0efb9902-75fb-4f9b-8966-8b70f61cd523","Type":"ContainerStarted","Data":"449d359e1c1fb34282ebc406f2b78d4a40af9deed60c816b26264312c5661985"} Dec 05 12:49:45 crc kubenswrapper[4809]: I1205 12:49:45.304128 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0efb9902-75fb-4f9b-8966-8b70f61cd523","Type":"ContainerStarted","Data":"805fdb3d02a645af3dc10dbb81c90cedb3090d166822c707fc5e65585e69e7b6"} Dec 05 12:49:45 crc kubenswrapper[4809]: I1205 12:49:45.818749 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:45 crc kubenswrapper[4809]: I1205 12:49:45.819165 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:49:46 crc kubenswrapper[4809]: I1205 12:49:46.332043 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0efb9902-75fb-4f9b-8966-8b70f61cd523","Type":"ContainerStarted","Data":"88031dd8c2e77d6e67b29a7a42b3ce784222f8829cf3ab7c0267fe66d1f71ff2"} Dec 05 12:49:46 crc kubenswrapper[4809]: I1205 12:49:46.337188 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"323c1d96-4a17-48b4-a300-57dd30a08cec","Type":"ContainerStarted","Data":"376e3e3c4b7a07eb02506e45149421f0fcd8a1acad04fd1cc9cff95c13db8d2d"} Dec 05 12:49:46 crc kubenswrapper[4809]: I1205 12:49:46.370073 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.370057266 podStartE2EDuration="3.370057266s" podCreationTimestamp="2025-12-05 12:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:49:46.368452093 +0000 UTC m=+6081.759428661" watchObservedRunningTime="2025-12-05 12:49:46.370057266 +0000 UTC m=+6081.761033824" Dec 05 12:49:46 crc kubenswrapper[4809]: I1205 12:49:46.401367 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.401340878 podStartE2EDuration="3.401340878s" podCreationTimestamp="2025-12-05 12:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:49:46.394993117 +0000 UTC m=+6081.785969675" watchObservedRunningTime="2025-12-05 12:49:46.401340878 +0000 UTC m=+6081.792317446" Dec 05 12:49:50 crc kubenswrapper[4809]: I1205 12:49:50.042367 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-lgwdk"] Dec 05 12:49:50 crc kubenswrapper[4809]: I1205 12:49:50.056136 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e34c-account-create-update-vffl5"] Dec 05 12:49:50 crc kubenswrapper[4809]: I1205 12:49:50.064421 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e34c-account-create-update-vffl5"] Dec 05 12:49:50 crc kubenswrapper[4809]: I1205 12:49:50.072649 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-lgwdk"] Dec 05 12:49:50 crc kubenswrapper[4809]: I1205 12:49:50.894614 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb1ffcc3-7e44-400b-a3c8-65071612f16f" path="/var/lib/kubelet/pods/cb1ffcc3-7e44-400b-a3c8-65071612f16f/volumes" Dec 05 12:49:50 crc kubenswrapper[4809]: I1205 12:49:50.895538 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc4bb97-a08f-40a3-b8f6-f014099c04cc" path="/var/lib/kubelet/pods/ccc4bb97-a08f-40a3-b8f6-f014099c04cc/volumes" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.622132 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.622724 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.660872 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.667506 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.736643 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.736723 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.766824 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:53 crc kubenswrapper[4809]: I1205 12:49:53.796955 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:54 crc kubenswrapper[4809]: I1205 12:49:54.419623 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:54 crc kubenswrapper[4809]: I1205 12:49:54.420245 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 12:49:54 crc kubenswrapper[4809]: I1205 12:49:54.420276 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:54 crc kubenswrapper[4809]: I1205 12:49:54.420286 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 12:49:55 crc kubenswrapper[4809]: I1205 12:49:55.204868 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-79969cdbc9-jcmsj" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 05 12:49:55 crc kubenswrapper[4809]: I1205 12:49:55.820712 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-855b9549c7-7gz4j" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.439679 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.439955 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.440216 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.440234 4809 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.475801 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.497292 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.497353 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 12:49:56 crc kubenswrapper[4809]: I1205 12:49:56.511805 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 12:50:03 crc kubenswrapper[4809]: I1205 12:50:03.037593 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-nxz6q"] Dec 05 12:50:03 crc kubenswrapper[4809]: I1205 12:50:03.052214 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-nxz6q"] Dec 05 12:50:04 crc kubenswrapper[4809]: I1205 12:50:04.884255 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67f36fae-3a5c-400e-afd9-1c3c33598973" path="/var/lib/kubelet/pods/67f36fae-3a5c-400e-afd9-1c3c33598973/volumes" Dec 05 12:50:07 crc kubenswrapper[4809]: I1205 12:50:07.003980 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:50:07 crc kubenswrapper[4809]: I1205 12:50:07.727560 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:50:08 crc kubenswrapper[4809]: I1205 12:50:08.721198 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:50:09 crc kubenswrapper[4809]: I1205 12:50:09.441665 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:50:09 crc kubenswrapper[4809]: I1205 12:50:09.493433 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79969cdbc9-jcmsj"] Dec 05 12:50:09 crc kubenswrapper[4809]: I1205 12:50:09.566261 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79969cdbc9-jcmsj" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon-log" containerID="cri-o://90ee1c66a6078532b65d785c6b746344291f4193843178889388821d0c5653e0" gracePeriod=30 Dec 05 12:50:09 crc kubenswrapper[4809]: I1205 12:50:09.566323 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79969cdbc9-jcmsj" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" containerID="cri-o://a1aca2ec9e8244d58893e3df45a37d7a1e87013edcaf1c0e09d5699b5157b059" gracePeriod=30 Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.575850 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.612749 4809 generic.go:334] "Generic (PLEG): container finished" podID="8556f0ed-27ba-4369-9a2b-325753529f06" containerID="a1aca2ec9e8244d58893e3df45a37d7a1e87013edcaf1c0e09d5699b5157b059" exitCode=0 Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.612813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79969cdbc9-jcmsj" event={"ID":"8556f0ed-27ba-4369-9a2b-325753529f06","Type":"ContainerDied","Data":"a1aca2ec9e8244d58893e3df45a37d7a1e87013edcaf1c0e09d5699b5157b059"} Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.617014 4809 generic.go:334] "Generic (PLEG): container finished" podID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerID="6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb" exitCode=137 Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.617037 4809 generic.go:334] "Generic (PLEG): container finished" podID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerID="ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86" exitCode=137 Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.617073 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-778c96f7-5qmgr" event={"ID":"dd3251ad-f094-4ab9-93b0-59c26aff5a40","Type":"ContainerDied","Data":"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb"} Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.617079 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-778c96f7-5qmgr" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.617096 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-778c96f7-5qmgr" event={"ID":"dd3251ad-f094-4ab9-93b0-59c26aff5a40","Type":"ContainerDied","Data":"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86"} Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.617106 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-778c96f7-5qmgr" event={"ID":"dd3251ad-f094-4ab9-93b0-59c26aff5a40","Type":"ContainerDied","Data":"014411af6d785d95a1f0dc8e3bc4de6e4ce967dad903634dfe0d9b456ddd446d"} Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.617121 4809 scope.go:117] "RemoveContainer" containerID="6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.686081 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-scripts\") pod \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.686270 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54pkq\" (UniqueName: \"kubernetes.io/projected/dd3251ad-f094-4ab9-93b0-59c26aff5a40-kube-api-access-54pkq\") pod \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.686320 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd3251ad-f094-4ab9-93b0-59c26aff5a40-logs\") pod \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.686348 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd3251ad-f094-4ab9-93b0-59c26aff5a40-horizon-secret-key\") pod \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.686380 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-config-data\") pod \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\" (UID: \"dd3251ad-f094-4ab9-93b0-59c26aff5a40\") " Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.687844 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd3251ad-f094-4ab9-93b0-59c26aff5a40-logs" (OuterVolumeSpecName: "logs") pod "dd3251ad-f094-4ab9-93b0-59c26aff5a40" (UID: "dd3251ad-f094-4ab9-93b0-59c26aff5a40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.692312 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3251ad-f094-4ab9-93b0-59c26aff5a40-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dd3251ad-f094-4ab9-93b0-59c26aff5a40" (UID: "dd3251ad-f094-4ab9-93b0-59c26aff5a40"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.692352 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd3251ad-f094-4ab9-93b0-59c26aff5a40-kube-api-access-54pkq" (OuterVolumeSpecName: "kube-api-access-54pkq") pod "dd3251ad-f094-4ab9-93b0-59c26aff5a40" (UID: "dd3251ad-f094-4ab9-93b0-59c26aff5a40"). InnerVolumeSpecName "kube-api-access-54pkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.713138 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-config-data" (OuterVolumeSpecName: "config-data") pod "dd3251ad-f094-4ab9-93b0-59c26aff5a40" (UID: "dd3251ad-f094-4ab9-93b0-59c26aff5a40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.717130 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-scripts" (OuterVolumeSpecName: "scripts") pod "dd3251ad-f094-4ab9-93b0-59c26aff5a40" (UID: "dd3251ad-f094-4ab9-93b0-59c26aff5a40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.790110 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.790149 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54pkq\" (UniqueName: \"kubernetes.io/projected/dd3251ad-f094-4ab9-93b0-59c26aff5a40-kube-api-access-54pkq\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.790161 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd3251ad-f094-4ab9-93b0-59c26aff5a40-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.790174 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd3251ad-f094-4ab9-93b0-59c26aff5a40-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.790185 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd3251ad-f094-4ab9-93b0-59c26aff5a40-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.802069 4809 scope.go:117] "RemoveContainer" containerID="ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.833512 4809 scope.go:117] "RemoveContainer" containerID="6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb" Dec 05 12:50:13 crc kubenswrapper[4809]: E1205 12:50:13.834444 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb\": container with ID starting with 6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb not found: ID does not exist" containerID="6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.834517 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb"} err="failed to get container status \"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb\": rpc error: code = NotFound desc = could not find container \"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb\": container with ID starting with 6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb not found: ID does not exist" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.834557 4809 scope.go:117] "RemoveContainer" containerID="ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86" Dec 05 12:50:13 crc kubenswrapper[4809]: E1205 12:50:13.835089 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86\": container with ID starting with ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86 not found: ID does not exist" containerID="ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.835137 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86"} err="failed to get container status \"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86\": rpc error: code = NotFound desc = could not find container \"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86\": container with ID starting with ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86 not found: ID does not exist" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.835166 4809 scope.go:117] "RemoveContainer" containerID="6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.836423 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb"} err="failed to get container status \"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb\": rpc error: code = NotFound desc = could not find container \"6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb\": container with ID starting with 6d402b96c511ae12de6448f396ffc2c7e915912f5cc2a1351af4dae4e233a2eb not found: ID does not exist" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.836475 4809 scope.go:117] "RemoveContainer" containerID="ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.836864 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86"} err="failed to get container status \"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86\": rpc error: code = NotFound desc = could not find container \"ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86\": container with ID starting with ee17429e1360ef8fa2aab9aad1acffbce4c94df5c197ed9ab6c988a6470aca86 not found: ID does not exist" Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.960173 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-778c96f7-5qmgr"] Dec 05 12:50:13 crc kubenswrapper[4809]: I1205 12:50:13.972119 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-778c96f7-5qmgr"] Dec 05 12:50:14 crc kubenswrapper[4809]: E1205 12:50:14.184802 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd3251ad_f094_4ab9_93b0_59c26aff5a40.slice\": RecentStats: unable to find data in memory cache]" Dec 05 12:50:14 crc kubenswrapper[4809]: I1205 12:50:14.893188 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" path="/var/lib/kubelet/pods/dd3251ad-f094-4ab9-93b0-59c26aff5a40/volumes" Dec 05 12:50:15 crc kubenswrapper[4809]: I1205 12:50:15.203804 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79969cdbc9-jcmsj" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.621000 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6db4cc6f8f-j562n"] Dec 05 12:50:16 crc kubenswrapper[4809]: E1205 12:50:16.621652 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.621665 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon" Dec 05 12:50:16 crc kubenswrapper[4809]: E1205 12:50:16.621694 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon-log" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.621700 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon-log" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.621940 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.621964 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3251ad-f094-4ab9-93b0-59c26aff5a40" containerName="horizon-log" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.623255 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.645410 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/89b8338a-fd8b-44cd-8a96-9863586f40ac-horizon-secret-key\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.645494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89b8338a-fd8b-44cd-8a96-9863586f40ac-config-data\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.645697 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbtp9\" (UniqueName: \"kubernetes.io/projected/89b8338a-fd8b-44cd-8a96-9863586f40ac-kube-api-access-zbtp9\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.645739 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89b8338a-fd8b-44cd-8a96-9863586f40ac-scripts\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.645772 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b8338a-fd8b-44cd-8a96-9863586f40ac-logs\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.649684 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6db4cc6f8f-j562n"] Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.748330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbtp9\" (UniqueName: \"kubernetes.io/projected/89b8338a-fd8b-44cd-8a96-9863586f40ac-kube-api-access-zbtp9\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.748423 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89b8338a-fd8b-44cd-8a96-9863586f40ac-scripts\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.748468 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b8338a-fd8b-44cd-8a96-9863586f40ac-logs\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.748578 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/89b8338a-fd8b-44cd-8a96-9863586f40ac-horizon-secret-key\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.748653 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89b8338a-fd8b-44cd-8a96-9863586f40ac-config-data\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.750335 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89b8338a-fd8b-44cd-8a96-9863586f40ac-logs\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.750747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89b8338a-fd8b-44cd-8a96-9863586f40ac-config-data\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.751024 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89b8338a-fd8b-44cd-8a96-9863586f40ac-scripts\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.757332 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/89b8338a-fd8b-44cd-8a96-9863586f40ac-horizon-secret-key\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.766835 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbtp9\" (UniqueName: \"kubernetes.io/projected/89b8338a-fd8b-44cd-8a96-9863586f40ac-kube-api-access-zbtp9\") pod \"horizon-6db4cc6f8f-j562n\" (UID: \"89b8338a-fd8b-44cd-8a96-9863586f40ac\") " pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:16 crc kubenswrapper[4809]: I1205 12:50:16.940919 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:17 crc kubenswrapper[4809]: I1205 12:50:17.503676 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6db4cc6f8f-j562n"] Dec 05 12:50:17 crc kubenswrapper[4809]: I1205 12:50:17.661783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db4cc6f8f-j562n" event={"ID":"89b8338a-fd8b-44cd-8a96-9863586f40ac","Type":"ContainerStarted","Data":"c9f3baba540259c249abcf03eed03c7391626ac6b5fcc4e9fd86d4f7db47d23f"} Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.110549 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-779pj"] Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.112403 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.130102 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-779pj"] Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.185485 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79fd2\" (UniqueName: \"kubernetes.io/projected/3a084734-6cf7-42ff-bdc8-21f5ad887112-kube-api-access-79fd2\") pod \"heat-db-create-779pj\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.189985 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a084734-6cf7-42ff-bdc8-21f5ad887112-operator-scripts\") pod \"heat-db-create-779pj\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.202275 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-6130-account-create-update-hnvrr"] Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.204136 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.207284 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.222406 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6130-account-create-update-hnvrr"] Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.292592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4adb3f3f-2746-42e1-bd82-6abebbe95479-operator-scripts\") pod \"heat-6130-account-create-update-hnvrr\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.292776 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a084734-6cf7-42ff-bdc8-21f5ad887112-operator-scripts\") pod \"heat-db-create-779pj\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.292827 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z25zl\" (UniqueName: \"kubernetes.io/projected/4adb3f3f-2746-42e1-bd82-6abebbe95479-kube-api-access-z25zl\") pod \"heat-6130-account-create-update-hnvrr\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.292925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79fd2\" (UniqueName: \"kubernetes.io/projected/3a084734-6cf7-42ff-bdc8-21f5ad887112-kube-api-access-79fd2\") pod \"heat-db-create-779pj\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.294072 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a084734-6cf7-42ff-bdc8-21f5ad887112-operator-scripts\") pod \"heat-db-create-779pj\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.311196 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79fd2\" (UniqueName: \"kubernetes.io/projected/3a084734-6cf7-42ff-bdc8-21f5ad887112-kube-api-access-79fd2\") pod \"heat-db-create-779pj\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.396444 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4adb3f3f-2746-42e1-bd82-6abebbe95479-operator-scripts\") pod \"heat-6130-account-create-update-hnvrr\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.397103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z25zl\" (UniqueName: \"kubernetes.io/projected/4adb3f3f-2746-42e1-bd82-6abebbe95479-kube-api-access-z25zl\") pod \"heat-6130-account-create-update-hnvrr\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.397244 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4adb3f3f-2746-42e1-bd82-6abebbe95479-operator-scripts\") pod \"heat-6130-account-create-update-hnvrr\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.413609 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z25zl\" (UniqueName: \"kubernetes.io/projected/4adb3f3f-2746-42e1-bd82-6abebbe95479-kube-api-access-z25zl\") pod \"heat-6130-account-create-update-hnvrr\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.478176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-779pj" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.523387 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.680897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db4cc6f8f-j562n" event={"ID":"89b8338a-fd8b-44cd-8a96-9863586f40ac","Type":"ContainerStarted","Data":"1d1959da07be04e67d4efd5871f972980ccb4d9fb65d24485882617a389a5bf8"} Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.681694 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db4cc6f8f-j562n" event={"ID":"89b8338a-fd8b-44cd-8a96-9863586f40ac","Type":"ContainerStarted","Data":"ca0ecf3829d3df0012e4114ee4a8659d31900ebd4f934dc242931973db09c492"} Dec 05 12:50:18 crc kubenswrapper[4809]: I1205 12:50:18.709450 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6db4cc6f8f-j562n" podStartSLOduration=2.709429029 podStartE2EDuration="2.709429029s" podCreationTimestamp="2025-12-05 12:50:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:50:18.702481212 +0000 UTC m=+6114.093457780" watchObservedRunningTime="2025-12-05 12:50:18.709429029 +0000 UTC m=+6114.100405587" Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.036685 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6130-account-create-update-hnvrr"] Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.091081 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-779pj"] Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.691133 4809 generic.go:334] "Generic (PLEG): container finished" podID="4adb3f3f-2746-42e1-bd82-6abebbe95479" containerID="e19164098b2dd8317026e51d4f477e6224d7897f83ec09c1044a8f5f6fc5d31c" exitCode=0 Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.691235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6130-account-create-update-hnvrr" event={"ID":"4adb3f3f-2746-42e1-bd82-6abebbe95479","Type":"ContainerDied","Data":"e19164098b2dd8317026e51d4f477e6224d7897f83ec09c1044a8f5f6fc5d31c"} Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.691291 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6130-account-create-update-hnvrr" event={"ID":"4adb3f3f-2746-42e1-bd82-6abebbe95479","Type":"ContainerStarted","Data":"113abc221df793cf9c8e058b13f3bf5fbe626f575ded6ac2c286a3423dc084c3"} Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.692897 4809 generic.go:334] "Generic (PLEG): container finished" podID="3a084734-6cf7-42ff-bdc8-21f5ad887112" containerID="2ac99a5bb22feed1a7518bb3b05d0b84b76c7f8f079c627af9d74f7e11fcfb45" exitCode=0 Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.693219 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-779pj" event={"ID":"3a084734-6cf7-42ff-bdc8-21f5ad887112","Type":"ContainerDied","Data":"2ac99a5bb22feed1a7518bb3b05d0b84b76c7f8f079c627af9d74f7e11fcfb45"} Dec 05 12:50:19 crc kubenswrapper[4809]: I1205 12:50:19.693289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-779pj" event={"ID":"3a084734-6cf7-42ff-bdc8-21f5ad887112","Type":"ContainerStarted","Data":"9cb89e33738747d409ce96da2aa1eb77911c84499937b136af25855e8532aa6f"} Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.182487 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.194250 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-779pj" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.257875 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79fd2\" (UniqueName: \"kubernetes.io/projected/3a084734-6cf7-42ff-bdc8-21f5ad887112-kube-api-access-79fd2\") pod \"3a084734-6cf7-42ff-bdc8-21f5ad887112\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.257932 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a084734-6cf7-42ff-bdc8-21f5ad887112-operator-scripts\") pod \"3a084734-6cf7-42ff-bdc8-21f5ad887112\" (UID: \"3a084734-6cf7-42ff-bdc8-21f5ad887112\") " Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.258022 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z25zl\" (UniqueName: \"kubernetes.io/projected/4adb3f3f-2746-42e1-bd82-6abebbe95479-kube-api-access-z25zl\") pod \"4adb3f3f-2746-42e1-bd82-6abebbe95479\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.258091 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4adb3f3f-2746-42e1-bd82-6abebbe95479-operator-scripts\") pod \"4adb3f3f-2746-42e1-bd82-6abebbe95479\" (UID: \"4adb3f3f-2746-42e1-bd82-6abebbe95479\") " Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.258696 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a084734-6cf7-42ff-bdc8-21f5ad887112-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a084734-6cf7-42ff-bdc8-21f5ad887112" (UID: "3a084734-6cf7-42ff-bdc8-21f5ad887112"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.258840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4adb3f3f-2746-42e1-bd82-6abebbe95479-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4adb3f3f-2746-42e1-bd82-6abebbe95479" (UID: "4adb3f3f-2746-42e1-bd82-6abebbe95479"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.264109 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4adb3f3f-2746-42e1-bd82-6abebbe95479-kube-api-access-z25zl" (OuterVolumeSpecName: "kube-api-access-z25zl") pod "4adb3f3f-2746-42e1-bd82-6abebbe95479" (UID: "4adb3f3f-2746-42e1-bd82-6abebbe95479"). InnerVolumeSpecName "kube-api-access-z25zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.264248 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a084734-6cf7-42ff-bdc8-21f5ad887112-kube-api-access-79fd2" (OuterVolumeSpecName: "kube-api-access-79fd2") pod "3a084734-6cf7-42ff-bdc8-21f5ad887112" (UID: "3a084734-6cf7-42ff-bdc8-21f5ad887112"). InnerVolumeSpecName "kube-api-access-79fd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.361330 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79fd2\" (UniqueName: \"kubernetes.io/projected/3a084734-6cf7-42ff-bdc8-21f5ad887112-kube-api-access-79fd2\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.361400 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a084734-6cf7-42ff-bdc8-21f5ad887112-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.361415 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z25zl\" (UniqueName: \"kubernetes.io/projected/4adb3f3f-2746-42e1-bd82-6abebbe95479-kube-api-access-z25zl\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.361429 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4adb3f3f-2746-42e1-bd82-6abebbe95479-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.716254 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-779pj" event={"ID":"3a084734-6cf7-42ff-bdc8-21f5ad887112","Type":"ContainerDied","Data":"9cb89e33738747d409ce96da2aa1eb77911c84499937b136af25855e8532aa6f"} Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.716303 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cb89e33738747d409ce96da2aa1eb77911c84499937b136af25855e8532aa6f" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.716363 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-779pj" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.721819 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6130-account-create-update-hnvrr" event={"ID":"4adb3f3f-2746-42e1-bd82-6abebbe95479","Type":"ContainerDied","Data":"113abc221df793cf9c8e058b13f3bf5fbe626f575ded6ac2c286a3423dc084c3"} Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.721872 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="113abc221df793cf9c8e058b13f3bf5fbe626f575ded6ac2c286a3423dc084c3" Dec 05 12:50:21 crc kubenswrapper[4809]: I1205 12:50:21.721850 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6130-account-create-update-hnvrr" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.407385 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-lgtr2"] Dec 05 12:50:23 crc kubenswrapper[4809]: E1205 12:50:23.408016 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4adb3f3f-2746-42e1-bd82-6abebbe95479" containerName="mariadb-account-create-update" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.408028 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4adb3f3f-2746-42e1-bd82-6abebbe95479" containerName="mariadb-account-create-update" Dec 05 12:50:23 crc kubenswrapper[4809]: E1205 12:50:23.408054 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a084734-6cf7-42ff-bdc8-21f5ad887112" containerName="mariadb-database-create" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.408061 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a084734-6cf7-42ff-bdc8-21f5ad887112" containerName="mariadb-database-create" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.408256 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4adb3f3f-2746-42e1-bd82-6abebbe95479" containerName="mariadb-account-create-update" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.408269 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a084734-6cf7-42ff-bdc8-21f5ad887112" containerName="mariadb-database-create" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.408913 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.411837 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.412338 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-lxg85" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.430283 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-lgtr2"] Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.606848 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-combined-ca-bundle\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.606889 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksczj\" (UniqueName: \"kubernetes.io/projected/95f30dfe-d44a-4316-b5e5-dd5193352cc8-kube-api-access-ksczj\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.607792 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-config-data\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.710115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-combined-ca-bundle\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.710171 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksczj\" (UniqueName: \"kubernetes.io/projected/95f30dfe-d44a-4316-b5e5-dd5193352cc8-kube-api-access-ksczj\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.710330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-config-data\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.716223 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-combined-ca-bundle\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.717085 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-config-data\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:23 crc kubenswrapper[4809]: I1205 12:50:23.734642 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksczj\" (UniqueName: \"kubernetes.io/projected/95f30dfe-d44a-4316-b5e5-dd5193352cc8-kube-api-access-ksczj\") pod \"heat-db-sync-lgtr2\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:24 crc kubenswrapper[4809]: I1205 12:50:24.033433 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:24 crc kubenswrapper[4809]: I1205 12:50:24.500070 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-lgtr2"] Dec 05 12:50:24 crc kubenswrapper[4809]: I1205 12:50:24.502258 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:50:24 crc kubenswrapper[4809]: I1205 12:50:24.781583 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lgtr2" event={"ID":"95f30dfe-d44a-4316-b5e5-dd5193352cc8","Type":"ContainerStarted","Data":"56efecaefb1155b1a99e42d7fae56d17d211fdc6f53ec3a57697fd96f4c83ee9"} Dec 05 12:50:25 crc kubenswrapper[4809]: I1205 12:50:25.205495 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79969cdbc9-jcmsj" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 05 12:50:26 crc kubenswrapper[4809]: I1205 12:50:26.941350 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:26 crc kubenswrapper[4809]: I1205 12:50:26.942876 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:30 crc kubenswrapper[4809]: I1205 12:50:30.840783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lgtr2" event={"ID":"95f30dfe-d44a-4316-b5e5-dd5193352cc8","Type":"ContainerStarted","Data":"f3d033246874cd8481c18d41ea7fd528ad21fdcd51a87ed24f7695a9a9098245"} Dec 05 12:50:30 crc kubenswrapper[4809]: I1205 12:50:30.868362 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-lgtr2" podStartSLOduration=2.304585434 podStartE2EDuration="7.868342813s" podCreationTimestamp="2025-12-05 12:50:23 +0000 UTC" firstStartedPulling="2025-12-05 12:50:24.502019637 +0000 UTC m=+6119.892996195" lastFinishedPulling="2025-12-05 12:50:30.065777016 +0000 UTC m=+6125.456753574" observedRunningTime="2025-12-05 12:50:30.854393608 +0000 UTC m=+6126.245370176" watchObservedRunningTime="2025-12-05 12:50:30.868342813 +0000 UTC m=+6126.259319381" Dec 05 12:50:32 crc kubenswrapper[4809]: I1205 12:50:32.866440 4809 generic.go:334] "Generic (PLEG): container finished" podID="95f30dfe-d44a-4316-b5e5-dd5193352cc8" containerID="f3d033246874cd8481c18d41ea7fd528ad21fdcd51a87ed24f7695a9a9098245" exitCode=0 Dec 05 12:50:32 crc kubenswrapper[4809]: I1205 12:50:32.866511 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lgtr2" event={"ID":"95f30dfe-d44a-4316-b5e5-dd5193352cc8","Type":"ContainerDied","Data":"f3d033246874cd8481c18d41ea7fd528ad21fdcd51a87ed24f7695a9a9098245"} Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.325565 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.431733 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksczj\" (UniqueName: \"kubernetes.io/projected/95f30dfe-d44a-4316-b5e5-dd5193352cc8-kube-api-access-ksczj\") pod \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.431833 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-combined-ca-bundle\") pod \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.431927 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-config-data\") pod \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\" (UID: \"95f30dfe-d44a-4316-b5e5-dd5193352cc8\") " Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.438176 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f30dfe-d44a-4316-b5e5-dd5193352cc8-kube-api-access-ksczj" (OuterVolumeSpecName: "kube-api-access-ksczj") pod "95f30dfe-d44a-4316-b5e5-dd5193352cc8" (UID: "95f30dfe-d44a-4316-b5e5-dd5193352cc8"). InnerVolumeSpecName "kube-api-access-ksczj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.492515 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95f30dfe-d44a-4316-b5e5-dd5193352cc8" (UID: "95f30dfe-d44a-4316-b5e5-dd5193352cc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.534232 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.534260 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksczj\" (UniqueName: \"kubernetes.io/projected/95f30dfe-d44a-4316-b5e5-dd5193352cc8-kube-api-access-ksczj\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.560935 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-config-data" (OuterVolumeSpecName: "config-data") pod "95f30dfe-d44a-4316-b5e5-dd5193352cc8" (UID: "95f30dfe-d44a-4316-b5e5-dd5193352cc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.636391 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f30dfe-d44a-4316-b5e5-dd5193352cc8-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.887269 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lgtr2" Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.916376 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lgtr2" event={"ID":"95f30dfe-d44a-4316-b5e5-dd5193352cc8","Type":"ContainerDied","Data":"56efecaefb1155b1a99e42d7fae56d17d211fdc6f53ec3a57697fd96f4c83ee9"} Dec 05 12:50:34 crc kubenswrapper[4809]: I1205 12:50:34.916422 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56efecaefb1155b1a99e42d7fae56d17d211fdc6f53ec3a57697fd96f4c83ee9" Dec 05 12:50:35 crc kubenswrapper[4809]: I1205 12:50:35.204061 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79969cdbc9-jcmsj" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.113:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8080: connect: connection refused" Dec 05 12:50:35 crc kubenswrapper[4809]: I1205 12:50:35.204185 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:50:35 crc kubenswrapper[4809]: I1205 12:50:35.992793 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-68cf595575-nttrk"] Dec 05 12:50:35 crc kubenswrapper[4809]: E1205 12:50:35.993487 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f30dfe-d44a-4316-b5e5-dd5193352cc8" containerName="heat-db-sync" Dec 05 12:50:35 crc kubenswrapper[4809]: I1205 12:50:35.993500 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f30dfe-d44a-4316-b5e5-dd5193352cc8" containerName="heat-db-sync" Dec 05 12:50:35 crc kubenswrapper[4809]: I1205 12:50:35.993752 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f30dfe-d44a-4316-b5e5-dd5193352cc8" containerName="heat-db-sync" Dec 05 12:50:35 crc kubenswrapper[4809]: I1205 12:50:35.994453 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.001310 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-lxg85" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.001592 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.001743 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.013027 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-68cf595575-nttrk"] Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.067863 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-combined-ca-bundle\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.067978 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwzh4\" (UniqueName: \"kubernetes.io/projected/39f3b59e-52c9-49a9-b63a-2129f74c2640-kube-api-access-gwzh4\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.068088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-config-data-custom\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.068125 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-config-data\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.075335 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-755878f84-9sw72"] Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.077192 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.081670 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.094119 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-755878f84-9sw72"] Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.141167 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-66d9fc74dc-mv9p7"] Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.148696 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.151613 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.162279 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-66d9fc74dc-mv9p7"] Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.169846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwzh4\" (UniqueName: \"kubernetes.io/projected/39f3b59e-52c9-49a9-b63a-2129f74c2640-kube-api-access-gwzh4\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.169947 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-config-data-custom\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.169975 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-config-data\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.170015 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w29k\" (UniqueName: \"kubernetes.io/projected/0b40e46c-2ed7-4277-8971-e366814ed23f-kube-api-access-9w29k\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.170055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-combined-ca-bundle\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.170088 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-config-data-custom\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.170112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-config-data\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.170131 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-combined-ca-bundle\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.179600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-config-data-custom\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.179702 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-combined-ca-bundle\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.180246 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f3b59e-52c9-49a9-b63a-2129f74c2640-config-data\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.205195 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwzh4\" (UniqueName: \"kubernetes.io/projected/39f3b59e-52c9-49a9-b63a-2129f74c2640-kube-api-access-gwzh4\") pod \"heat-engine-68cf595575-nttrk\" (UID: \"39f3b59e-52c9-49a9-b63a-2129f74c2640\") " pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.271792 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkxfq\" (UniqueName: \"kubernetes.io/projected/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-kube-api-access-tkxfq\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.271835 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-config-data-custom\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.271867 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-combined-ca-bundle\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.271892 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w29k\" (UniqueName: \"kubernetes.io/projected/0b40e46c-2ed7-4277-8971-e366814ed23f-kube-api-access-9w29k\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.271928 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-config-data\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.271945 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-combined-ca-bundle\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.271977 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-config-data-custom\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.272002 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-config-data\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.276205 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-config-data-custom\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.277087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-config-data\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.280717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b40e46c-2ed7-4277-8971-e366814ed23f-combined-ca-bundle\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.293167 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w29k\" (UniqueName: \"kubernetes.io/projected/0b40e46c-2ed7-4277-8971-e366814ed23f-kube-api-access-9w29k\") pod \"heat-cfnapi-755878f84-9sw72\" (UID: \"0b40e46c-2ed7-4277-8971-e366814ed23f\") " pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.333588 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.382506 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-combined-ca-bundle\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.387688 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-config-data\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.388182 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkxfq\" (UniqueName: \"kubernetes.io/projected/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-kube-api-access-tkxfq\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.388215 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-config-data-custom\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.393834 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-config-data-custom\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.394004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-combined-ca-bundle\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.395295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-config-data\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.402325 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.419081 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkxfq\" (UniqueName: \"kubernetes.io/projected/6ebcec3a-3b2f-4b08-aa54-1701d35d272f-kube-api-access-tkxfq\") pod \"heat-api-66d9fc74dc-mv9p7\" (UID: \"6ebcec3a-3b2f-4b08-aa54-1701d35d272f\") " pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.465158 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.937939 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-68cf595575-nttrk"] Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.944093 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6db4cc6f8f-j562n" podUID="89b8338a-fd8b-44cd-8a96-9863586f40ac" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.117:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8080: connect: connection refused" Dec 05 12:50:36 crc kubenswrapper[4809]: I1205 12:50:36.954531 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-755878f84-9sw72"] Dec 05 12:50:37 crc kubenswrapper[4809]: I1205 12:50:37.095304 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-66d9fc74dc-mv9p7"] Dec 05 12:50:37 crc kubenswrapper[4809]: I1205 12:50:37.935993 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-755878f84-9sw72" event={"ID":"0b40e46c-2ed7-4277-8971-e366814ed23f","Type":"ContainerStarted","Data":"92405a417252ceada3206570a78c14024f3435167abe23e52cb63834e70c4382"} Dec 05 12:50:37 crc kubenswrapper[4809]: I1205 12:50:37.937398 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-66d9fc74dc-mv9p7" event={"ID":"6ebcec3a-3b2f-4b08-aa54-1701d35d272f","Type":"ContainerStarted","Data":"a4264fcce5d12fc66d3ad0ba115d8b3dc5482918d8e774e466fdcf19e40d1394"} Dec 05 12:50:37 crc kubenswrapper[4809]: I1205 12:50:37.939151 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-68cf595575-nttrk" event={"ID":"39f3b59e-52c9-49a9-b63a-2129f74c2640","Type":"ContainerStarted","Data":"ddc4d5d40d6d0223c41fabdd5f1dfbfc5c1baf961fe8acacde9394e3eec23d69"} Dec 05 12:50:37 crc kubenswrapper[4809]: I1205 12:50:37.939199 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-68cf595575-nttrk" event={"ID":"39f3b59e-52c9-49a9-b63a-2129f74c2640","Type":"ContainerStarted","Data":"b1bf4180c8eb6b22d89b97d63900a8782fcadc9ec6f08220445457f67cf14c10"} Dec 05 12:50:37 crc kubenswrapper[4809]: I1205 12:50:37.939241 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:50:37 crc kubenswrapper[4809]: I1205 12:50:37.961867 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-68cf595575-nttrk" podStartSLOduration=2.961848037 podStartE2EDuration="2.961848037s" podCreationTimestamp="2025-12-05 12:50:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:50:37.958289542 +0000 UTC m=+6133.349266120" watchObservedRunningTime="2025-12-05 12:50:37.961848037 +0000 UTC m=+6133.352824595" Dec 05 12:50:39 crc kubenswrapper[4809]: I1205 12:50:39.961369 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-66d9fc74dc-mv9p7" event={"ID":"6ebcec3a-3b2f-4b08-aa54-1701d35d272f","Type":"ContainerStarted","Data":"7b21b4b45505cbcc29d278e28e7e2e590cd056d8c304698351ef9a5e9d7086e7"} Dec 05 12:50:39 crc kubenswrapper[4809]: I1205 12:50:39.963000 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:39 crc kubenswrapper[4809]: I1205 12:50:39.966095 4809 generic.go:334] "Generic (PLEG): container finished" podID="8556f0ed-27ba-4369-9a2b-325753529f06" containerID="90ee1c66a6078532b65d785c6b746344291f4193843178889388821d0c5653e0" exitCode=137 Dec 05 12:50:39 crc kubenswrapper[4809]: I1205 12:50:39.966161 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79969cdbc9-jcmsj" event={"ID":"8556f0ed-27ba-4369-9a2b-325753529f06","Type":"ContainerDied","Data":"90ee1c66a6078532b65d785c6b746344291f4193843178889388821d0c5653e0"} Dec 05 12:50:39 crc kubenswrapper[4809]: I1205 12:50:39.968220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-755878f84-9sw72" event={"ID":"0b40e46c-2ed7-4277-8971-e366814ed23f","Type":"ContainerStarted","Data":"5ded900d63fa85052fb2bc525598b9958fbab1145ca9f5001ce47aa739b81604"} Dec 05 12:50:39 crc kubenswrapper[4809]: I1205 12:50:39.968435 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:39 crc kubenswrapper[4809]: I1205 12:50:39.989588 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-66d9fc74dc-mv9p7" podStartSLOduration=2.087156229 podStartE2EDuration="3.989569362s" podCreationTimestamp="2025-12-05 12:50:36 +0000 UTC" firstStartedPulling="2025-12-05 12:50:37.098900065 +0000 UTC m=+6132.489876623" lastFinishedPulling="2025-12-05 12:50:39.001313188 +0000 UTC m=+6134.392289756" observedRunningTime="2025-12-05 12:50:39.982984565 +0000 UTC m=+6135.373961123" watchObservedRunningTime="2025-12-05 12:50:39.989569362 +0000 UTC m=+6135.380545920" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.007846 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-755878f84-9sw72" podStartSLOduration=1.9390048119999999 podStartE2EDuration="4.007815923s" podCreationTimestamp="2025-12-05 12:50:36 +0000 UTC" firstStartedPulling="2025-12-05 12:50:36.939483815 +0000 UTC m=+6132.330460373" lastFinishedPulling="2025-12-05 12:50:39.008294936 +0000 UTC m=+6134.399271484" observedRunningTime="2025-12-05 12:50:40.004746521 +0000 UTC m=+6135.395723079" watchObservedRunningTime="2025-12-05 12:50:40.007815923 +0000 UTC m=+6135.398792491" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.078506 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.183331 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8556f0ed-27ba-4369-9a2b-325753529f06-horizon-secret-key\") pod \"8556f0ed-27ba-4369-9a2b-325753529f06\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.183437 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9fws\" (UniqueName: \"kubernetes.io/projected/8556f0ed-27ba-4369-9a2b-325753529f06-kube-api-access-d9fws\") pod \"8556f0ed-27ba-4369-9a2b-325753529f06\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.183520 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-config-data\") pod \"8556f0ed-27ba-4369-9a2b-325753529f06\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.183571 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8556f0ed-27ba-4369-9a2b-325753529f06-logs\") pod \"8556f0ed-27ba-4369-9a2b-325753529f06\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.183669 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-scripts\") pod \"8556f0ed-27ba-4369-9a2b-325753529f06\" (UID: \"8556f0ed-27ba-4369-9a2b-325753529f06\") " Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.193013 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8556f0ed-27ba-4369-9a2b-325753529f06-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8556f0ed-27ba-4369-9a2b-325753529f06" (UID: "8556f0ed-27ba-4369-9a2b-325753529f06"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.193285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8556f0ed-27ba-4369-9a2b-325753529f06-logs" (OuterVolumeSpecName: "logs") pod "8556f0ed-27ba-4369-9a2b-325753529f06" (UID: "8556f0ed-27ba-4369-9a2b-325753529f06"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.206745 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8556f0ed-27ba-4369-9a2b-325753529f06-kube-api-access-d9fws" (OuterVolumeSpecName: "kube-api-access-d9fws") pod "8556f0ed-27ba-4369-9a2b-325753529f06" (UID: "8556f0ed-27ba-4369-9a2b-325753529f06"). InnerVolumeSpecName "kube-api-access-d9fws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.253379 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-config-data" (OuterVolumeSpecName: "config-data") pod "8556f0ed-27ba-4369-9a2b-325753529f06" (UID: "8556f0ed-27ba-4369-9a2b-325753529f06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.286142 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8556f0ed-27ba-4369-9a2b-325753529f06-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.286180 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8556f0ed-27ba-4369-9a2b-325753529f06-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.286191 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9fws\" (UniqueName: \"kubernetes.io/projected/8556f0ed-27ba-4369-9a2b-325753529f06-kube-api-access-d9fws\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.286201 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.301186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-scripts" (OuterVolumeSpecName: "scripts") pod "8556f0ed-27ba-4369-9a2b-325753529f06" (UID: "8556f0ed-27ba-4369-9a2b-325753529f06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.387969 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8556f0ed-27ba-4369-9a2b-325753529f06-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.982630 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79969cdbc9-jcmsj" event={"ID":"8556f0ed-27ba-4369-9a2b-325753529f06","Type":"ContainerDied","Data":"bc12d0c1f1c0d29fd10297c7d3f3e89b51e667a1a34851c87da413bf0ab9f4b9"} Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.983013 4809 scope.go:117] "RemoveContainer" containerID="a1aca2ec9e8244d58893e3df45a37d7a1e87013edcaf1c0e09d5699b5157b059" Dec 05 12:50:40 crc kubenswrapper[4809]: I1205 12:50:40.982802 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79969cdbc9-jcmsj" Dec 05 12:50:41 crc kubenswrapper[4809]: I1205 12:50:41.005325 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79969cdbc9-jcmsj"] Dec 05 12:50:41 crc kubenswrapper[4809]: I1205 12:50:41.016889 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79969cdbc9-jcmsj"] Dec 05 12:50:41 crc kubenswrapper[4809]: I1205 12:50:41.176985 4809 scope.go:117] "RemoveContainer" containerID="90ee1c66a6078532b65d785c6b746344291f4193843178889388821d0c5653e0" Dec 05 12:50:42 crc kubenswrapper[4809]: I1205 12:50:42.220919 4809 scope.go:117] "RemoveContainer" containerID="853062e39ea30216ee76ee1fada3f2dad4b363851250da8dc37db47099fafcee" Dec 05 12:50:42 crc kubenswrapper[4809]: I1205 12:50:42.554090 4809 scope.go:117] "RemoveContainer" containerID="3b4c71e457b2f63fa20300186a49d3267b0cb88c80c27fa2d12936932574c81a" Dec 05 12:50:42 crc kubenswrapper[4809]: I1205 12:50:42.867405 4809 scope.go:117] "RemoveContainer" containerID="f3bfe3273ad663d02342cfa66f8aeb7369345b1e3287267477fc080ecfdf51c6" Dec 05 12:50:42 crc kubenswrapper[4809]: I1205 12:50:42.884301 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" path="/var/lib/kubelet/pods/8556f0ed-27ba-4369-9a2b-325753529f06/volumes" Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.037078 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-85d5-account-create-update-x9nkx"] Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.047216 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.047562 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.049335 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-cpwzd"] Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.066811 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-85d5-account-create-update-x9nkx"] Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.078345 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-cpwzd"] Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.891504 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad6aadd-94ef-435a-a013-d04f6fde42d8" path="/var/lib/kubelet/pods/7ad6aadd-94ef-435a-a013-d04f6fde42d8/volumes" Dec 05 12:50:44 crc kubenswrapper[4809]: I1205 12:50:44.892655 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b101ef7e-9934-4bb5-90d5-211a8a42128d" path="/var/lib/kubelet/pods/b101ef7e-9934-4bb5-90d5-211a8a42128d/volumes" Dec 05 12:50:47 crc kubenswrapper[4809]: I1205 12:50:47.917584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-66d9fc74dc-mv9p7" Dec 05 12:50:47 crc kubenswrapper[4809]: I1205 12:50:47.949470 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-755878f84-9sw72" Dec 05 12:50:49 crc kubenswrapper[4809]: I1205 12:50:49.143232 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:50 crc kubenswrapper[4809]: I1205 12:50:50.864575 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6db4cc6f8f-j562n" Dec 05 12:50:51 crc kubenswrapper[4809]: I1205 12:50:51.164774 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-855b9549c7-7gz4j"] Dec 05 12:50:51 crc kubenswrapper[4809]: I1205 12:50:51.165014 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-855b9549c7-7gz4j" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon-log" containerID="cri-o://69be850a99b8866ddc464c1264d39c7f17f9f7f430a0d7e677d274f2b5baf3c9" gracePeriod=30 Dec 05 12:50:51 crc kubenswrapper[4809]: I1205 12:50:51.165499 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-855b9549c7-7gz4j" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" containerID="cri-o://410fa6e02707d6e1a8e040478cafffa70737c869e9fe716ccbb6a2b162601460" gracePeriod=30 Dec 05 12:50:52 crc kubenswrapper[4809]: I1205 12:50:52.062325 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-n6m5r"] Dec 05 12:50:52 crc kubenswrapper[4809]: I1205 12:50:52.075519 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-n6m5r"] Dec 05 12:50:52 crc kubenswrapper[4809]: I1205 12:50:52.885030 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b944ba07-ed9e-4304-bcb7-8c54e80e27d9" path="/var/lib/kubelet/pods/b944ba07-ed9e-4304-bcb7-8c54e80e27d9/volumes" Dec 05 12:50:55 crc kubenswrapper[4809]: I1205 12:50:55.142888 4809 generic.go:334] "Generic (PLEG): container finished" podID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerID="410fa6e02707d6e1a8e040478cafffa70737c869e9fe716ccbb6a2b162601460" exitCode=0 Dec 05 12:50:55 crc kubenswrapper[4809]: I1205 12:50:55.142927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855b9549c7-7gz4j" event={"ID":"2373ada0-4330-4520-a4a8-975f8be6a4f3","Type":"ContainerDied","Data":"410fa6e02707d6e1a8e040478cafffa70737c869e9fe716ccbb6a2b162601460"} Dec 05 12:50:55 crc kubenswrapper[4809]: I1205 12:50:55.820212 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-855b9549c7-7gz4j" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Dec 05 12:50:56 crc kubenswrapper[4809]: I1205 12:50:56.362814 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-68cf595575-nttrk" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.529861 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd"] Dec 05 12:51:05 crc kubenswrapper[4809]: E1205 12:51:05.530885 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon-log" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.530901 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon-log" Dec 05 12:51:05 crc kubenswrapper[4809]: E1205 12:51:05.530935 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.530940 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.531137 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon-log" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.531152 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8556f0ed-27ba-4369-9a2b-325753529f06" containerName="horizon" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.532572 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.534574 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.540607 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd"] Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.724872 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.726537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgh7t\" (UniqueName: \"kubernetes.io/projected/59537950-4298-4f29-8513-d246da286b29-kube-api-access-rgh7t\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.727050 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.819134 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-855b9549c7-7gz4j" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.829596 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.829744 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgh7t\" (UniqueName: \"kubernetes.io/projected/59537950-4298-4f29-8513-d246da286b29-kube-api-access-rgh7t\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.830066 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.830531 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.830719 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.857783 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgh7t\" (UniqueName: \"kubernetes.io/projected/59537950-4298-4f29-8513-d246da286b29-kube-api-access-rgh7t\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:05 crc kubenswrapper[4809]: I1205 12:51:05.868113 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:06 crc kubenswrapper[4809]: I1205 12:51:06.568043 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd"] Dec 05 12:51:06 crc kubenswrapper[4809]: W1205 12:51:06.571749 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59537950_4298_4f29_8513_d246da286b29.slice/crio-0df79bf15e9962b650c8a42e976ae5bf59a9a78a658e7147a5d3094418258960 WatchSource:0}: Error finding container 0df79bf15e9962b650c8a42e976ae5bf59a9a78a658e7147a5d3094418258960: Status 404 returned error can't find the container with id 0df79bf15e9962b650c8a42e976ae5bf59a9a78a658e7147a5d3094418258960 Dec 05 12:51:07 crc kubenswrapper[4809]: I1205 12:51:07.310301 4809 generic.go:334] "Generic (PLEG): container finished" podID="59537950-4298-4f29-8513-d246da286b29" containerID="aca10f208e2572d4ecac32fc846089b3d87defca235bab7bd7d3bcc628027f90" exitCode=0 Dec 05 12:51:07 crc kubenswrapper[4809]: I1205 12:51:07.310399 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" event={"ID":"59537950-4298-4f29-8513-d246da286b29","Type":"ContainerDied","Data":"aca10f208e2572d4ecac32fc846089b3d87defca235bab7bd7d3bcc628027f90"} Dec 05 12:51:07 crc kubenswrapper[4809]: I1205 12:51:07.310947 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" event={"ID":"59537950-4298-4f29-8513-d246da286b29","Type":"ContainerStarted","Data":"0df79bf15e9962b650c8a42e976ae5bf59a9a78a658e7147a5d3094418258960"} Dec 05 12:51:07 crc kubenswrapper[4809]: I1205 12:51:07.876767 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lmfpm"] Dec 05 12:51:07 crc kubenswrapper[4809]: I1205 12:51:07.880272 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:07 crc kubenswrapper[4809]: I1205 12:51:07.900008 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lmfpm"] Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.078515 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmm6x\" (UniqueName: \"kubernetes.io/projected/4e2adec0-4ec2-4feb-8827-214ecc6eefda-kube-api-access-kmm6x\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.079226 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-catalog-content\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.079401 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-utilities\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.184957 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-catalog-content\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.185064 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-utilities\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.185137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmm6x\" (UniqueName: \"kubernetes.io/projected/4e2adec0-4ec2-4feb-8827-214ecc6eefda-kube-api-access-kmm6x\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.185516 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-catalog-content\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.185567 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-utilities\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.206749 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmm6x\" (UniqueName: \"kubernetes.io/projected/4e2adec0-4ec2-4feb-8827-214ecc6eefda-kube-api-access-kmm6x\") pod \"redhat-operators-lmfpm\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.210517 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:08 crc kubenswrapper[4809]: I1205 12:51:08.895167 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lmfpm"] Dec 05 12:51:08 crc kubenswrapper[4809]: W1205 12:51:08.905072 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e2adec0_4ec2_4feb_8827_214ecc6eefda.slice/crio-83657b2cea8ee525884d9a385eedd241dcd1a053fe690aba594b3c7618fe0c45 WatchSource:0}: Error finding container 83657b2cea8ee525884d9a385eedd241dcd1a053fe690aba594b3c7618fe0c45: Status 404 returned error can't find the container with id 83657b2cea8ee525884d9a385eedd241dcd1a053fe690aba594b3c7618fe0c45 Dec 05 12:51:09 crc kubenswrapper[4809]: I1205 12:51:09.332752 4809 generic.go:334] "Generic (PLEG): container finished" podID="59537950-4298-4f29-8513-d246da286b29" containerID="2ecfe9a8db3828ea910cf03e082daa747494d86c8ff139906d59f7843c2cea0a" exitCode=0 Dec 05 12:51:09 crc kubenswrapper[4809]: I1205 12:51:09.332856 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" event={"ID":"59537950-4298-4f29-8513-d246da286b29","Type":"ContainerDied","Data":"2ecfe9a8db3828ea910cf03e082daa747494d86c8ff139906d59f7843c2cea0a"} Dec 05 12:51:09 crc kubenswrapper[4809]: I1205 12:51:09.334935 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerID="55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883" exitCode=0 Dec 05 12:51:09 crc kubenswrapper[4809]: I1205 12:51:09.334978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmfpm" event={"ID":"4e2adec0-4ec2-4feb-8827-214ecc6eefda","Type":"ContainerDied","Data":"55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883"} Dec 05 12:51:09 crc kubenswrapper[4809]: I1205 12:51:09.335006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmfpm" event={"ID":"4e2adec0-4ec2-4feb-8827-214ecc6eefda","Type":"ContainerStarted","Data":"83657b2cea8ee525884d9a385eedd241dcd1a053fe690aba594b3c7618fe0c45"} Dec 05 12:51:10 crc kubenswrapper[4809]: I1205 12:51:10.347402 4809 generic.go:334] "Generic (PLEG): container finished" podID="59537950-4298-4f29-8513-d246da286b29" containerID="fd501e15eeeaafd1d75c23f66745f5d61dc8c60216852ebfc30e3bf0b78b86ce" exitCode=0 Dec 05 12:51:10 crc kubenswrapper[4809]: I1205 12:51:10.347684 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" event={"ID":"59537950-4298-4f29-8513-d246da286b29","Type":"ContainerDied","Data":"fd501e15eeeaafd1d75c23f66745f5d61dc8c60216852ebfc30e3bf0b78b86ce"} Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.359159 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmfpm" event={"ID":"4e2adec0-4ec2-4feb-8827-214ecc6eefda","Type":"ContainerStarted","Data":"566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6"} Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.728370 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.795162 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-util\") pod \"59537950-4298-4f29-8513-d246da286b29\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.795223 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-bundle\") pod \"59537950-4298-4f29-8513-d246da286b29\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.795402 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgh7t\" (UniqueName: \"kubernetes.io/projected/59537950-4298-4f29-8513-d246da286b29-kube-api-access-rgh7t\") pod \"59537950-4298-4f29-8513-d246da286b29\" (UID: \"59537950-4298-4f29-8513-d246da286b29\") " Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.799083 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-bundle" (OuterVolumeSpecName: "bundle") pod "59537950-4298-4f29-8513-d246da286b29" (UID: "59537950-4298-4f29-8513-d246da286b29"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.802323 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59537950-4298-4f29-8513-d246da286b29-kube-api-access-rgh7t" (OuterVolumeSpecName: "kube-api-access-rgh7t") pod "59537950-4298-4f29-8513-d246da286b29" (UID: "59537950-4298-4f29-8513-d246da286b29"). InnerVolumeSpecName "kube-api-access-rgh7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.812883 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-util" (OuterVolumeSpecName: "util") pod "59537950-4298-4f29-8513-d246da286b29" (UID: "59537950-4298-4f29-8513-d246da286b29"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.897868 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgh7t\" (UniqueName: \"kubernetes.io/projected/59537950-4298-4f29-8513-d246da286b29-kube-api-access-rgh7t\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.897915 4809 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-util\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:11 crc kubenswrapper[4809]: I1205 12:51:11.897948 4809 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59537950-4298-4f29-8513-d246da286b29-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:12 crc kubenswrapper[4809]: I1205 12:51:12.398783 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" event={"ID":"59537950-4298-4f29-8513-d246da286b29","Type":"ContainerDied","Data":"0df79bf15e9962b650c8a42e976ae5bf59a9a78a658e7147a5d3094418258960"} Dec 05 12:51:12 crc kubenswrapper[4809]: I1205 12:51:12.398815 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd" Dec 05 12:51:12 crc kubenswrapper[4809]: I1205 12:51:12.398832 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0df79bf15e9962b650c8a42e976ae5bf59a9a78a658e7147a5d3094418258960" Dec 05 12:51:13 crc kubenswrapper[4809]: I1205 12:51:13.411862 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerID="566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6" exitCode=0 Dec 05 12:51:13 crc kubenswrapper[4809]: I1205 12:51:13.411891 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmfpm" event={"ID":"4e2adec0-4ec2-4feb-8827-214ecc6eefda","Type":"ContainerDied","Data":"566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6"} Dec 05 12:51:14 crc kubenswrapper[4809]: I1205 12:51:14.046998 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:51:14 crc kubenswrapper[4809]: I1205 12:51:14.047047 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:51:14 crc kubenswrapper[4809]: I1205 12:51:14.422418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmfpm" event={"ID":"4e2adec0-4ec2-4feb-8827-214ecc6eefda","Type":"ContainerStarted","Data":"2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff"} Dec 05 12:51:14 crc kubenswrapper[4809]: I1205 12:51:14.448690 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lmfpm" podStartSLOduration=2.717621916 podStartE2EDuration="7.448672697s" podCreationTimestamp="2025-12-05 12:51:07 +0000 UTC" firstStartedPulling="2025-12-05 12:51:09.33607082 +0000 UTC m=+6164.727047368" lastFinishedPulling="2025-12-05 12:51:14.067121591 +0000 UTC m=+6169.458098149" observedRunningTime="2025-12-05 12:51:14.441395571 +0000 UTC m=+6169.832372139" watchObservedRunningTime="2025-12-05 12:51:14.448672697 +0000 UTC m=+6169.839649255" Dec 05 12:51:15 crc kubenswrapper[4809]: I1205 12:51:15.819551 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-855b9549c7-7gz4j" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Dec 05 12:51:15 crc kubenswrapper[4809]: I1205 12:51:15.819712 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:51:18 crc kubenswrapper[4809]: I1205 12:51:18.211844 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:18 crc kubenswrapper[4809]: I1205 12:51:18.212154 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:19 crc kubenswrapper[4809]: I1205 12:51:19.296884 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lmfpm" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="registry-server" probeResult="failure" output=< Dec 05 12:51:19 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 12:51:19 crc kubenswrapper[4809]: > Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.490512 4809 generic.go:334] "Generic (PLEG): container finished" podID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerID="69be850a99b8866ddc464c1264d39c7f17f9f7f430a0d7e677d274f2b5baf3c9" exitCode=137 Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.490595 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855b9549c7-7gz4j" event={"ID":"2373ada0-4330-4520-a4a8-975f8be6a4f3","Type":"ContainerDied","Data":"69be850a99b8866ddc464c1264d39c7f17f9f7f430a0d7e677d274f2b5baf3c9"} Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.665751 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.800768 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2373ada0-4330-4520-a4a8-975f8be6a4f3-horizon-secret-key\") pod \"2373ada0-4330-4520-a4a8-975f8be6a4f3\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.800876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-scripts\") pod \"2373ada0-4330-4520-a4a8-975f8be6a4f3\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.801015 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-config-data\") pod \"2373ada0-4330-4520-a4a8-975f8be6a4f3\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.801034 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps2kn\" (UniqueName: \"kubernetes.io/projected/2373ada0-4330-4520-a4a8-975f8be6a4f3-kube-api-access-ps2kn\") pod \"2373ada0-4330-4520-a4a8-975f8be6a4f3\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.801096 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373ada0-4330-4520-a4a8-975f8be6a4f3-logs\") pod \"2373ada0-4330-4520-a4a8-975f8be6a4f3\" (UID: \"2373ada0-4330-4520-a4a8-975f8be6a4f3\") " Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.801834 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2373ada0-4330-4520-a4a8-975f8be6a4f3-logs" (OuterVolumeSpecName: "logs") pod "2373ada0-4330-4520-a4a8-975f8be6a4f3" (UID: "2373ada0-4330-4520-a4a8-975f8be6a4f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.817033 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2373ada0-4330-4520-a4a8-975f8be6a4f3-kube-api-access-ps2kn" (OuterVolumeSpecName: "kube-api-access-ps2kn") pod "2373ada0-4330-4520-a4a8-975f8be6a4f3" (UID: "2373ada0-4330-4520-a4a8-975f8be6a4f3"). InnerVolumeSpecName "kube-api-access-ps2kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.840855 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2373ada0-4330-4520-a4a8-975f8be6a4f3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2373ada0-4330-4520-a4a8-975f8be6a4f3" (UID: "2373ada0-4330-4520-a4a8-975f8be6a4f3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.843212 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-scripts" (OuterVolumeSpecName: "scripts") pod "2373ada0-4330-4520-a4a8-975f8be6a4f3" (UID: "2373ada0-4330-4520-a4a8-975f8be6a4f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.860760 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-config-data" (OuterVolumeSpecName: "config-data") pod "2373ada0-4330-4520-a4a8-975f8be6a4f3" (UID: "2373ada0-4330-4520-a4a8-975f8be6a4f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.903948 4809 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2373ada0-4330-4520-a4a8-975f8be6a4f3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.903990 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.904002 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2373ada0-4330-4520-a4a8-975f8be6a4f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.904011 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps2kn\" (UniqueName: \"kubernetes.io/projected/2373ada0-4330-4520-a4a8-975f8be6a4f3-kube-api-access-ps2kn\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:21 crc kubenswrapper[4809]: I1205 12:51:21.904022 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373ada0-4330-4520-a4a8-975f8be6a4f3-logs\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.045162 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wgztn"] Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.060430 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wgztn"] Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.072786 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6087-account-create-update-pbc2b"] Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.083706 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6087-account-create-update-pbc2b"] Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.500174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855b9549c7-7gz4j" event={"ID":"2373ada0-4330-4520-a4a8-975f8be6a4f3","Type":"ContainerDied","Data":"6b91b3fe82dcfb2989f193135e2f1f8a9bc66aded2b8b4f3e05e71d92586c8e7"} Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.500410 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855b9549c7-7gz4j" Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.500479 4809 scope.go:117] "RemoveContainer" containerID="410fa6e02707d6e1a8e040478cafffa70737c869e9fe716ccbb6a2b162601460" Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.558221 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-855b9549c7-7gz4j"] Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.566728 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-855b9549c7-7gz4j"] Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.672740 4809 scope.go:117] "RemoveContainer" containerID="69be850a99b8866ddc464c1264d39c7f17f9f7f430a0d7e677d274f2b5baf3c9" Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.883931 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" path="/var/lib/kubelet/pods/2373ada0-4330-4520-a4a8-975f8be6a4f3/volumes" Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.884624 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3103bf31-c6f8-494f-a75e-deb3cea6d915" path="/var/lib/kubelet/pods/3103bf31-c6f8-494f-a75e-deb3cea6d915/volumes" Dec 05 12:51:22 crc kubenswrapper[4809]: I1205 12:51:22.885202 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b95a840-4da9-4162-bf78-a89b0c0ce3ba" path="/var/lib/kubelet/pods/4b95a840-4da9-4162-bf78-a89b0c0ce3ba/volumes" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.819376 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m"] Dec 05 12:51:24 crc kubenswrapper[4809]: E1205 12:51:24.821329 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59537950-4298-4f29-8513-d246da286b29" containerName="pull" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821350 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="59537950-4298-4f29-8513-d246da286b29" containerName="pull" Dec 05 12:51:24 crc kubenswrapper[4809]: E1205 12:51:24.821369 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59537950-4298-4f29-8513-d246da286b29" containerName="util" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821378 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="59537950-4298-4f29-8513-d246da286b29" containerName="util" Dec 05 12:51:24 crc kubenswrapper[4809]: E1205 12:51:24.821400 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59537950-4298-4f29-8513-d246da286b29" containerName="extract" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821410 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="59537950-4298-4f29-8513-d246da286b29" containerName="extract" Dec 05 12:51:24 crc kubenswrapper[4809]: E1205 12:51:24.821423 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821431 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" Dec 05 12:51:24 crc kubenswrapper[4809]: E1205 12:51:24.821446 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon-log" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821473 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon-log" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821788 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon-log" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821830 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2373ada0-4330-4520-a4a8-975f8be6a4f3" containerName="horizon" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.821843 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="59537950-4298-4f29-8513-d246da286b29" containerName="extract" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.822788 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.828129 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.828403 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.828518 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-7z7x8" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.831089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m"] Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.933133 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf"] Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.934523 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.936553 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.936644 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-tfwdz" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.952953 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf"] Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.960732 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx"] Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.961898 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:24 crc kubenswrapper[4809]: I1205 12:51:24.963508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brnjp\" (UniqueName: \"kubernetes.io/projected/2412d41f-89ce-4153-9c3a-c281e9ed6406-kube-api-access-brnjp\") pod \"obo-prometheus-operator-668cf9dfbb-srt4m\" (UID: \"2412d41f-89ce-4153-9c3a-c281e9ed6406\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.026038 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx"] Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.073858 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf\" (UID: \"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.074057 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf\" (UID: \"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.074157 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brnjp\" (UniqueName: \"kubernetes.io/projected/2412d41f-89ce-4153-9c3a-c281e9ed6406-kube-api-access-brnjp\") pod \"obo-prometheus-operator-668cf9dfbb-srt4m\" (UID: \"2412d41f-89ce-4153-9c3a-c281e9ed6406\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.074211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60278de2-7102-42b7-b617-567279d19923-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx\" (UID: \"60278de2-7102-42b7-b617-567279d19923\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.074295 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60278de2-7102-42b7-b617-567279d19923-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx\" (UID: \"60278de2-7102-42b7-b617-567279d19923\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.091942 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.102778 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.128408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brnjp\" (UniqueName: \"kubernetes.io/projected/2412d41f-89ce-4153-9c3a-c281e9ed6406-kube-api-access-brnjp\") pod \"obo-prometheus-operator-668cf9dfbb-srt4m\" (UID: \"2412d41f-89ce-4153-9c3a-c281e9ed6406\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.157127 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5zfk8"] Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.158808 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.169752 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-7sc64" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.169942 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-7z7x8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.170091 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.170726 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.177022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf\" (UID: \"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.177141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60278de2-7102-42b7-b617-567279d19923-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx\" (UID: \"60278de2-7102-42b7-b617-567279d19923\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.177166 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60278de2-7102-42b7-b617-567279d19923-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx\" (UID: \"60278de2-7102-42b7-b617-567279d19923\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.177316 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf\" (UID: \"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.189124 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60278de2-7102-42b7-b617-567279d19923-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx\" (UID: \"60278de2-7102-42b7-b617-567279d19923\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.189292 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf\" (UID: \"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.198301 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf\" (UID: \"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.204555 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60278de2-7102-42b7-b617-567279d19923-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx\" (UID: \"60278de2-7102-42b7-b617-567279d19923\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.205378 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5zfk8"] Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.266236 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.280787 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e48c383-a2d1-4a9c-a17f-1d6b157ef997-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5zfk8\" (UID: \"5e48c383-a2d1-4a9c-a17f-1d6b157ef997\") " pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.280859 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8g5d\" (UniqueName: \"kubernetes.io/projected/5e48c383-a2d1-4a9c-a17f-1d6b157ef997-kube-api-access-m8g5d\") pod \"observability-operator-d8bb48f5d-5zfk8\" (UID: \"5e48c383-a2d1-4a9c-a17f-1d6b157ef997\") " pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.315532 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.403483 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e48c383-a2d1-4a9c-a17f-1d6b157ef997-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5zfk8\" (UID: \"5e48c383-a2d1-4a9c-a17f-1d6b157ef997\") " pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.403835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8g5d\" (UniqueName: \"kubernetes.io/projected/5e48c383-a2d1-4a9c-a17f-1d6b157ef997-kube-api-access-m8g5d\") pod \"observability-operator-d8bb48f5d-5zfk8\" (UID: \"5e48c383-a2d1-4a9c-a17f-1d6b157ef997\") " pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.422035 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e48c383-a2d1-4a9c-a17f-1d6b157ef997-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5zfk8\" (UID: \"5e48c383-a2d1-4a9c-a17f-1d6b157ef997\") " pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.425490 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8g5d\" (UniqueName: \"kubernetes.io/projected/5e48c383-a2d1-4a9c-a17f-1d6b157ef997-kube-api-access-m8g5d\") pod \"observability-operator-d8bb48f5d-5zfk8\" (UID: \"5e48c383-a2d1-4a9c-a17f-1d6b157ef997\") " pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.493712 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hqwhc"] Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.497731 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.503124 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-brd9n" Dec 05 12:51:25 crc kubenswrapper[4809]: I1205 12:51:25.506493 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hqwhc"] Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.613874 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg5lx\" (UniqueName: \"kubernetes.io/projected/bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63-kube-api-access-kg5lx\") pod \"perses-operator-5446b9c989-hqwhc\" (UID: \"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63\") " pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.613984 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63-openshift-service-ca\") pod \"perses-operator-5446b9c989-hqwhc\" (UID: \"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63\") " pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.616237 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.716942 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63-openshift-service-ca\") pod \"perses-operator-5446b9c989-hqwhc\" (UID: \"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63\") " pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.717076 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg5lx\" (UniqueName: \"kubernetes.io/projected/bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63-kube-api-access-kg5lx\") pod \"perses-operator-5446b9c989-hqwhc\" (UID: \"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63\") " pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.718912 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63-openshift-service-ca\") pod \"perses-operator-5446b9c989-hqwhc\" (UID: \"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63\") " pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.744086 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg5lx\" (UniqueName: \"kubernetes.io/projected/bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63-kube-api-access-kg5lx\") pod \"perses-operator-5446b9c989-hqwhc\" (UID: \"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63\") " pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.838072 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.901870 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf"] Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:25.940490 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m"] Dec 05 12:51:26 crc kubenswrapper[4809]: W1205 12:51:26.017474 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2412d41f_89ce_4153_9c3a_c281e9ed6406.slice/crio-b29fe7968d573bae9a97e76be1d09689b1dac5bb6a5f9426722caa34c5f1d4ac WatchSource:0}: Error finding container b29fe7968d573bae9a97e76be1d09689b1dac5bb6a5f9426722caa34c5f1d4ac: Status 404 returned error can't find the container with id b29fe7968d573bae9a97e76be1d09689b1dac5bb6a5f9426722caa34c5f1d4ac Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:26.059589 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx"] Dec 05 12:51:26 crc kubenswrapper[4809]: W1205 12:51:26.093456 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60278de2_7102_42b7_b617_567279d19923.slice/crio-3885159c7410dc14fb74490880c854d5e14d7be8899c46191344553c11a345f0 WatchSource:0}: Error finding container 3885159c7410dc14fb74490880c854d5e14d7be8899c46191344553c11a345f0: Status 404 returned error can't find the container with id 3885159c7410dc14fb74490880c854d5e14d7be8899c46191344553c11a345f0 Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:26.548938 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" event={"ID":"2412d41f-89ce-4153-9c3a-c281e9ed6406","Type":"ContainerStarted","Data":"b29fe7968d573bae9a97e76be1d09689b1dac5bb6a5f9426722caa34c5f1d4ac"} Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:26.549948 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" event={"ID":"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea","Type":"ContainerStarted","Data":"04f15b7caf522bca900a11e3f82bd3fa046da1e5971e715e8db5d6ec5bca1db5"} Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:26.550867 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" event={"ID":"60278de2-7102-42b7-b617-567279d19923","Type":"ContainerStarted","Data":"3885159c7410dc14fb74490880c854d5e14d7be8899c46191344553c11a345f0"} Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:26.762884 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hqwhc"] Dec 05 12:51:26 crc kubenswrapper[4809]: W1205 12:51:26.763971 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd2d4a74_3f06_4ad0_8387_a6fe2e8f2d63.slice/crio-2b42af44e6ae3c054a8c28344debbbf388d9793603d2f4d7895528db92adef27 WatchSource:0}: Error finding container 2b42af44e6ae3c054a8c28344debbbf388d9793603d2f4d7895528db92adef27: Status 404 returned error can't find the container with id 2b42af44e6ae3c054a8c28344debbbf388d9793603d2f4d7895528db92adef27 Dec 05 12:51:26 crc kubenswrapper[4809]: I1205 12:51:26.770681 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5zfk8"] Dec 05 12:51:26 crc kubenswrapper[4809]: W1205 12:51:26.780460 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e48c383_a2d1_4a9c_a17f_1d6b157ef997.slice/crio-d3a8601dc6f6ee42e810d3f85812c144831761a9697e6ed5ce8a318a7c998a60 WatchSource:0}: Error finding container d3a8601dc6f6ee42e810d3f85812c144831761a9697e6ed5ce8a318a7c998a60: Status 404 returned error can't find the container with id d3a8601dc6f6ee42e810d3f85812c144831761a9697e6ed5ce8a318a7c998a60 Dec 05 12:51:27 crc kubenswrapper[4809]: I1205 12:51:27.028324 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-7r8q4"] Dec 05 12:51:27 crc kubenswrapper[4809]: I1205 12:51:27.036305 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-7r8q4"] Dec 05 12:51:27 crc kubenswrapper[4809]: I1205 12:51:27.561108 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-hqwhc" event={"ID":"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63","Type":"ContainerStarted","Data":"2b42af44e6ae3c054a8c28344debbbf388d9793603d2f4d7895528db92adef27"} Dec 05 12:51:27 crc kubenswrapper[4809]: I1205 12:51:27.563731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" event={"ID":"5e48c383-a2d1-4a9c-a17f-1d6b157ef997","Type":"ContainerStarted","Data":"d3a8601dc6f6ee42e810d3f85812c144831761a9697e6ed5ce8a318a7c998a60"} Dec 05 12:51:28 crc kubenswrapper[4809]: I1205 12:51:28.885693 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325aafe3-ade7-4fbb-b031-53e17b3e6422" path="/var/lib/kubelet/pods/325aafe3-ade7-4fbb-b031-53e17b3e6422/volumes" Dec 05 12:51:29 crc kubenswrapper[4809]: I1205 12:51:29.374231 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lmfpm" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="registry-server" probeResult="failure" output=< Dec 05 12:51:29 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 12:51:29 crc kubenswrapper[4809]: > Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.682005 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" event={"ID":"60278de2-7102-42b7-b617-567279d19923","Type":"ContainerStarted","Data":"c2be4ccc0afd73375b4e583dc260cf2581a3a110d924b3c59b89dc92e31a620d"} Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.686464 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" event={"ID":"5e48c383-a2d1-4a9c-a17f-1d6b157ef997","Type":"ContainerStarted","Data":"b55dfa310a2f6cb69ffad9e132359490d940c04f82120e141c7a53e5463c1055"} Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.686590 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.688765 4809 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-5zfk8 container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.129:8081/healthz\": dial tcp 10.217.1.129:8081: connect: connection refused" start-of-body= Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.688833 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" podUID="5e48c383-a2d1-4a9c-a17f-1d6b157ef997" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.129:8081/healthz\": dial tcp 10.217.1.129:8081: connect: connection refused" Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.693414 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" event={"ID":"a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea","Type":"ContainerStarted","Data":"a116e7dc258e98530feb9a34f8b74788ae40278e99696dff5de12d5365257acd"} Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.697344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-hqwhc" event={"ID":"bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63","Type":"ContainerStarted","Data":"27896a5869f60f45b0f251721cb48c45696a356df9871ca66a2286e0b1ca8487"} Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.698221 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.755190 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx" podStartSLOduration=2.858010586 podStartE2EDuration="11.755169979s" podCreationTimestamp="2025-12-05 12:51:24 +0000 UTC" firstStartedPulling="2025-12-05 12:51:26.0988565 +0000 UTC m=+6181.489833058" lastFinishedPulling="2025-12-05 12:51:34.996015893 +0000 UTC m=+6190.386992451" observedRunningTime="2025-12-05 12:51:35.725282105 +0000 UTC m=+6191.116258663" watchObservedRunningTime="2025-12-05 12:51:35.755169979 +0000 UTC m=+6191.146146537" Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.756664 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf" podStartSLOduration=2.736442265 podStartE2EDuration="11.756657779s" podCreationTimestamp="2025-12-05 12:51:24 +0000 UTC" firstStartedPulling="2025-12-05 12:51:25.976261852 +0000 UTC m=+6181.367238410" lastFinishedPulling="2025-12-05 12:51:34.996477366 +0000 UTC m=+6190.387453924" observedRunningTime="2025-12-05 12:51:35.746050563 +0000 UTC m=+6191.137027141" watchObservedRunningTime="2025-12-05 12:51:35.756657779 +0000 UTC m=+6191.147634337" Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.783137 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" podStartSLOduration=2.21051065 podStartE2EDuration="10.783106351s" podCreationTimestamp="2025-12-05 12:51:25 +0000 UTC" firstStartedPulling="2025-12-05 12:51:26.78496522 +0000 UTC m=+6182.175941778" lastFinishedPulling="2025-12-05 12:51:35.357560921 +0000 UTC m=+6190.748537479" observedRunningTime="2025-12-05 12:51:35.777027067 +0000 UTC m=+6191.168003635" watchObservedRunningTime="2025-12-05 12:51:35.783106351 +0000 UTC m=+6191.174082909" Dec 05 12:51:35 crc kubenswrapper[4809]: I1205 12:51:35.847475 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-hqwhc" podStartSLOduration=2.621073407 podStartE2EDuration="10.847454882s" podCreationTimestamp="2025-12-05 12:51:25 +0000 UTC" firstStartedPulling="2025-12-05 12:51:26.767698326 +0000 UTC m=+6182.158674884" lastFinishedPulling="2025-12-05 12:51:34.994079801 +0000 UTC m=+6190.385056359" observedRunningTime="2025-12-05 12:51:35.838053059 +0000 UTC m=+6191.229029617" watchObservedRunningTime="2025-12-05 12:51:35.847454882 +0000 UTC m=+6191.238431470" Dec 05 12:51:36 crc kubenswrapper[4809]: I1205 12:51:36.710730 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" event={"ID":"2412d41f-89ce-4153-9c3a-c281e9ed6406","Type":"ContainerStarted","Data":"3d87323d1d1bd72537779a4e3b12d3f5e40396817c9406dab764243d0fd66d18"} Dec 05 12:51:36 crc kubenswrapper[4809]: I1205 12:51:36.715201 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-5zfk8" Dec 05 12:51:36 crc kubenswrapper[4809]: I1205 12:51:36.731105 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-srt4m" podStartSLOduration=3.756419307 podStartE2EDuration="12.731079575s" podCreationTimestamp="2025-12-05 12:51:24 +0000 UTC" firstStartedPulling="2025-12-05 12:51:26.033389999 +0000 UTC m=+6181.424366557" lastFinishedPulling="2025-12-05 12:51:35.008050267 +0000 UTC m=+6190.399026825" observedRunningTime="2025-12-05 12:51:36.729416031 +0000 UTC m=+6192.120392599" watchObservedRunningTime="2025-12-05 12:51:36.731079575 +0000 UTC m=+6192.122056133" Dec 05 12:51:38 crc kubenswrapper[4809]: I1205 12:51:38.275521 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:38 crc kubenswrapper[4809]: I1205 12:51:38.324831 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.090751 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lmfpm"] Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.091230 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lmfpm" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="registry-server" containerID="cri-o://2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff" gracePeriod=2 Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.568357 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.658785 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-utilities\") pod \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.658905 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmm6x\" (UniqueName: \"kubernetes.io/projected/4e2adec0-4ec2-4feb-8827-214ecc6eefda-kube-api-access-kmm6x\") pod \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.658937 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-catalog-content\") pod \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\" (UID: \"4e2adec0-4ec2-4feb-8827-214ecc6eefda\") " Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.659548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-utilities" (OuterVolumeSpecName: "utilities") pod "4e2adec0-4ec2-4feb-8827-214ecc6eefda" (UID: "4e2adec0-4ec2-4feb-8827-214ecc6eefda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.664279 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e2adec0-4ec2-4feb-8827-214ecc6eefda-kube-api-access-kmm6x" (OuterVolumeSpecName: "kube-api-access-kmm6x") pod "4e2adec0-4ec2-4feb-8827-214ecc6eefda" (UID: "4e2adec0-4ec2-4feb-8827-214ecc6eefda"). InnerVolumeSpecName "kube-api-access-kmm6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.754435 4809 generic.go:334] "Generic (PLEG): container finished" podID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerID="2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff" exitCode=0 Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.754507 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmfpm" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.754527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmfpm" event={"ID":"4e2adec0-4ec2-4feb-8827-214ecc6eefda","Type":"ContainerDied","Data":"2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff"} Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.755027 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmfpm" event={"ID":"4e2adec0-4ec2-4feb-8827-214ecc6eefda","Type":"ContainerDied","Data":"83657b2cea8ee525884d9a385eedd241dcd1a053fe690aba594b3c7618fe0c45"} Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.755070 4809 scope.go:117] "RemoveContainer" containerID="2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.760335 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.760364 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmm6x\" (UniqueName: \"kubernetes.io/projected/4e2adec0-4ec2-4feb-8827-214ecc6eefda-kube-api-access-kmm6x\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.780796 4809 scope.go:117] "RemoveContainer" containerID="566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.801985 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e2adec0-4ec2-4feb-8827-214ecc6eefda" (UID: "4e2adec0-4ec2-4feb-8827-214ecc6eefda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.841871 4809 scope.go:117] "RemoveContainer" containerID="55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.863257 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2adec0-4ec2-4feb-8827-214ecc6eefda-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.913790 4809 scope.go:117] "RemoveContainer" containerID="2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff" Dec 05 12:51:40 crc kubenswrapper[4809]: E1205 12:51:40.921859 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff\": container with ID starting with 2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff not found: ID does not exist" containerID="2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.921921 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff"} err="failed to get container status \"2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff\": rpc error: code = NotFound desc = could not find container \"2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff\": container with ID starting with 2ee5461a0dd8075f4ed74e7f2d60aaa02e142daab2689b0afc2931859d559bff not found: ID does not exist" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.921964 4809 scope.go:117] "RemoveContainer" containerID="566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6" Dec 05 12:51:40 crc kubenswrapper[4809]: E1205 12:51:40.924185 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6\": container with ID starting with 566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6 not found: ID does not exist" containerID="566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.924220 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6"} err="failed to get container status \"566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6\": rpc error: code = NotFound desc = could not find container \"566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6\": container with ID starting with 566f85690f4030e73e288bcf20093320d07812fee9e16b864d31ea29fcf383f6 not found: ID does not exist" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.924245 4809 scope.go:117] "RemoveContainer" containerID="55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883" Dec 05 12:51:40 crc kubenswrapper[4809]: E1205 12:51:40.924675 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883\": container with ID starting with 55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883 not found: ID does not exist" containerID="55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883" Dec 05 12:51:40 crc kubenswrapper[4809]: I1205 12:51:40.924699 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883"} err="failed to get container status \"55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883\": rpc error: code = NotFound desc = could not find container \"55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883\": container with ID starting with 55bcd6b900378475b6da8a456618cb45a1208e46745bd9126b2573e46c82f883 not found: ID does not exist" Dec 05 12:51:41 crc kubenswrapper[4809]: I1205 12:51:41.086960 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lmfpm"] Dec 05 12:51:41 crc kubenswrapper[4809]: I1205 12:51:41.098210 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lmfpm"] Dec 05 12:51:42 crc kubenswrapper[4809]: I1205 12:51:42.887409 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" path="/var/lib/kubelet/pods/4e2adec0-4ec2-4feb-8827-214ecc6eefda/volumes" Dec 05 12:51:43 crc kubenswrapper[4809]: I1205 12:51:43.019022 4809 scope.go:117] "RemoveContainer" containerID="6a95f60470fe38859395be6af8241e50657cff321bf12ca416c9af8b109948f5" Dec 05 12:51:43 crc kubenswrapper[4809]: I1205 12:51:43.058861 4809 scope.go:117] "RemoveContainer" containerID="aa34681b2e890c1ddab1802193a81d2104d2513559aedcb9cb9471041a924ee0" Dec 05 12:51:43 crc kubenswrapper[4809]: I1205 12:51:43.087990 4809 scope.go:117] "RemoveContainer" containerID="6a3c444356bd95b62b53d5c7987521329b1cd06773c88ece1cee34b84341a42e" Dec 05 12:51:43 crc kubenswrapper[4809]: I1205 12:51:43.138011 4809 scope.go:117] "RemoveContainer" containerID="92b430e4e8024df5ac6b8285abe1bae3e637daeb0d04f4cc6058496cf0999da9" Dec 05 12:51:43 crc kubenswrapper[4809]: I1205 12:51:43.181920 4809 scope.go:117] "RemoveContainer" containerID="c2b6990d10915b0b0956cf8b9636b9187b3d650df6998b93c6e7e54222a7d4dc" Dec 05 12:51:43 crc kubenswrapper[4809]: I1205 12:51:43.238271 4809 scope.go:117] "RemoveContainer" containerID="029e738f53db4fb600d138c70628dbfb950fc25e99085270e213d435231b9298" Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.046449 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.046823 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.046878 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.048162 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e0ee3fefa7029bd441f660ece6169c58fb6143e423ed0039f9d6ad8d0c48340"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.048232 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://7e0ee3fefa7029bd441f660ece6169c58fb6143e423ed0039f9d6ad8d0c48340" gracePeriod=600 Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.804534 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="7e0ee3fefa7029bd441f660ece6169c58fb6143e423ed0039f9d6ad8d0c48340" exitCode=0 Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.804584 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"7e0ee3fefa7029bd441f660ece6169c58fb6143e423ed0039f9d6ad8d0c48340"} Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.805066 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391"} Dec 05 12:51:44 crc kubenswrapper[4809]: I1205 12:51:44.805088 4809 scope.go:117] "RemoveContainer" containerID="70cbd474bce681c2eb577d5fc1e7469a05195dc5f5ecd7db02f99d0aceb29b31" Dec 05 12:51:45 crc kubenswrapper[4809]: I1205 12:51:45.841619 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-hqwhc" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.455279 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.456034 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="73acb568-95c4-4d30-81bd-1056b887ea65" containerName="openstackclient" containerID="cri-o://9c6e66048455719a05e49fd6068d11524a5db88f7e8bbc27f4952110fcfee101" gracePeriod=2 Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.478465 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.497318 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: E1205 12:51:48.497762 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73acb568-95c4-4d30-81bd-1056b887ea65" containerName="openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.497779 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="73acb568-95c4-4d30-81bd-1056b887ea65" containerName="openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: E1205 12:51:48.497830 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="extract-utilities" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.497837 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="extract-utilities" Dec 05 12:51:48 crc kubenswrapper[4809]: E1205 12:51:48.497857 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="registry-server" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.497865 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="registry-server" Dec 05 12:51:48 crc kubenswrapper[4809]: E1205 12:51:48.497887 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="extract-content" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.497892 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="extract-content" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.498073 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2adec0-4ec2-4feb-8827-214ecc6eefda" containerName="registry-server" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.498092 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="73acb568-95c4-4d30-81bd-1056b887ea65" containerName="openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.498789 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.504868 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="73acb568-95c4-4d30-81bd-1056b887ea65" podUID="66c408e0-327a-4d11-b06d-c85e778107ae" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.511809 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.511940 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k97pc\" (UniqueName: \"kubernetes.io/projected/66c408e0-327a-4d11-b06d-c85e778107ae-kube-api-access-k97pc\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.511964 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.558006 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.613673 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: E1205 12:51:48.614523 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-k97pc openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="66c408e0-327a-4d11-b06d-c85e778107ae" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.614957 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k97pc\" (UniqueName: \"kubernetes.io/projected/66c408e0-327a-4d11-b06d-c85e778107ae-kube-api-access-k97pc\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.615009 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.615096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.615969 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: E1205 12:51:48.619198 4809 projected.go:194] Error preparing data for projected volume kube-api-access-k97pc for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (66c408e0-327a-4d11-b06d-c85e778107ae) does not match the UID in record. The object might have been deleted and then recreated Dec 05 12:51:48 crc kubenswrapper[4809]: E1205 12:51:48.619254 4809 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66c408e0-327a-4d11-b06d-c85e778107ae-kube-api-access-k97pc podName:66c408e0-327a-4d11-b06d-c85e778107ae nodeName:}" failed. No retries permitted until 2025-12-05 12:51:49.119236223 +0000 UTC m=+6204.510212781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-k97pc" (UniqueName: "kubernetes.io/projected/66c408e0-327a-4d11-b06d-c85e778107ae-kube-api-access-k97pc") pod "openstackclient" (UID: "66c408e0-327a-4d11-b06d-c85e778107ae") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (66c408e0-327a-4d11-b06d-c85e778107ae) does not match the UID in record. The object might have been deleted and then recreated Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.633623 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.644723 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.662218 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.663604 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.671543 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.778162 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.779581 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.781816 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-x66d2" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.795961 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.821541 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svjft\" (UniqueName: \"kubernetes.io/projected/c49ded6e-3eb8-44f1-9125-5938485a644a-kube-api-access-svjft\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.821675 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c49ded6e-3eb8-44f1-9125-5938485a644a-openstack-config-secret\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.821813 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c49ded6e-3eb8-44f1-9125-5938485a644a-openstack-config\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.844899 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.860744 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="66c408e0-327a-4d11-b06d-c85e778107ae" podUID="c49ded6e-3eb8-44f1-9125-5938485a644a" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.872581 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.925797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c49ded6e-3eb8-44f1-9125-5938485a644a-openstack-config\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.926092 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svjft\" (UniqueName: \"kubernetes.io/projected/c49ded6e-3eb8-44f1-9125-5938485a644a-kube-api-access-svjft\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.926213 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c49ded6e-3eb8-44f1-9125-5938485a644a-openstack-config-secret\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.926311 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lk97\" (UniqueName: \"kubernetes.io/projected/810023b7-c0cb-4c82-9d79-387562b0b0d1-kube-api-access-2lk97\") pod \"kube-state-metrics-0\" (UID: \"810023b7-c0cb-4c82-9d79-387562b0b0d1\") " pod="openstack/kube-state-metrics-0" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.926717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c49ded6e-3eb8-44f1-9125-5938485a644a-openstack-config\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.960155 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c49ded6e-3eb8-44f1-9125-5938485a644a-openstack-config-secret\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:48 crc kubenswrapper[4809]: I1205 12:51:48.983239 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svjft\" (UniqueName: \"kubernetes.io/projected/c49ded6e-3eb8-44f1-9125-5938485a644a-kube-api-access-svjft\") pod \"openstackclient\" (UID: \"c49ded6e-3eb8-44f1-9125-5938485a644a\") " pod="openstack/openstackclient" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.030937 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config\") pod \"66c408e0-327a-4d11-b06d-c85e778107ae\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.031221 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config-secret\") pod \"66c408e0-327a-4d11-b06d-c85e778107ae\" (UID: \"66c408e0-327a-4d11-b06d-c85e778107ae\") " Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.031460 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "66c408e0-327a-4d11-b06d-c85e778107ae" (UID: "66c408e0-327a-4d11-b06d-c85e778107ae"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.031746 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lk97\" (UniqueName: \"kubernetes.io/projected/810023b7-c0cb-4c82-9d79-387562b0b0d1-kube-api-access-2lk97\") pod \"kube-state-metrics-0\" (UID: \"810023b7-c0cb-4c82-9d79-387562b0b0d1\") " pod="openstack/kube-state-metrics-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.031853 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.031868 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k97pc\" (UniqueName: \"kubernetes.io/projected/66c408e0-327a-4d11-b06d-c85e778107ae-kube-api-access-k97pc\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.039685 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.043775 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "66c408e0-327a-4d11-b06d-c85e778107ae" (UID: "66c408e0-327a-4d11-b06d-c85e778107ae"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.116330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lk97\" (UniqueName: \"kubernetes.io/projected/810023b7-c0cb-4c82-9d79-387562b0b0d1-kube-api-access-2lk97\") pod \"kube-state-metrics-0\" (UID: \"810023b7-c0cb-4c82-9d79-387562b0b0d1\") " pod="openstack/kube-state-metrics-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.133368 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66c408e0-327a-4d11-b06d-c85e778107ae-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.402140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.849080 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.851719 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.859681 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.859867 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.859964 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.860058 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.860156 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-8qzpx" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.869762 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.935265 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.966550 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="66c408e0-327a-4d11-b06d-c85e778107ae" podUID="c49ded6e-3eb8-44f1-9125-5938485a644a" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.982273 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.982361 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndbsw\" (UniqueName: \"kubernetes.io/projected/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-kube-api-access-ndbsw\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.982434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.982510 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.982589 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.982670 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:49 crc kubenswrapper[4809]: I1205 12:51:49.982711 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.084890 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.085419 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.085530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.085600 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndbsw\" (UniqueName: \"kubernetes.io/projected/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-kube-api-access-ndbsw\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.085787 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.085905 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.086020 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.087923 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.093244 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.095059 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.098457 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.098832 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.102151 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.115278 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndbsw\" (UniqueName: \"kubernetes.io/projected/374aeffe-f715-4f6e-af5b-4ff7f547d9c5-kube-api-access-ndbsw\") pod \"alertmanager-metric-storage-0\" (UID: \"374aeffe-f715-4f6e-af5b-4ff7f547d9c5\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.177955 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.199618 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.413905 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.417832 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.422427 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.422606 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.422765 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.427910 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.428120 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-kxlnl" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.433144 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.459675 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.581017 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.639918 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0381cceb-e34d-4a34-bfb0-8673a52237dc-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.640016 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0381cceb-e34d-4a34-bfb0-8673a52237dc-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.640097 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k77b\" (UniqueName: \"kubernetes.io/projected/0381cceb-e34d-4a34-bfb0-8673a52237dc-kube-api-access-9k77b\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.640131 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.640223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-abeea079-7b02-4587-9a62-d3fd391b5253\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-abeea079-7b02-4587-9a62-d3fd391b5253\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.640258 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-config\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.640358 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.640385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0381cceb-e34d-4a34-bfb0-8673a52237dc-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.741456 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.741795 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-abeea079-7b02-4587-9a62-d3fd391b5253\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-abeea079-7b02-4587-9a62-d3fd391b5253\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.741829 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-config\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.741862 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.741883 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0381cceb-e34d-4a34-bfb0-8673a52237dc-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.741920 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0381cceb-e34d-4a34-bfb0-8673a52237dc-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.741955 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0381cceb-e34d-4a34-bfb0-8673a52237dc-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.742009 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k77b\" (UniqueName: \"kubernetes.io/projected/0381cceb-e34d-4a34-bfb0-8673a52237dc-kube-api-access-9k77b\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.744565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0381cceb-e34d-4a34-bfb0-8673a52237dc-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.746903 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.753593 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-config\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.753653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0381cceb-e34d-4a34-bfb0-8673a52237dc-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.753793 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0381cceb-e34d-4a34-bfb0-8673a52237dc-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.753911 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0381cceb-e34d-4a34-bfb0-8673a52237dc-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.754155 4809 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.754179 4809 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-abeea079-7b02-4587-9a62-d3fd391b5253\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-abeea079-7b02-4587-9a62-d3fd391b5253\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/64d87c146cf223ebe1a147ebd1bb134fa1fc638bd8c0a0c095e209d92491ead3/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.765578 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k77b\" (UniqueName: \"kubernetes.io/projected/0381cceb-e34d-4a34-bfb0-8673a52237dc-kube-api-access-9k77b\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.854275 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-abeea079-7b02-4587-9a62-d3fd391b5253\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-abeea079-7b02-4587-9a62-d3fd391b5253\") pod \"prometheus-metric-storage-0\" (UID: \"0381cceb-e34d-4a34-bfb0-8673a52237dc\") " pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.919263 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c408e0-327a-4d11-b06d-c85e778107ae" path="/var/lib/kubelet/pods/66c408e0-327a-4d11-b06d-c85e778107ae/volumes" Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.946795 4809 generic.go:334] "Generic (PLEG): container finished" podID="73acb568-95c4-4d30-81bd-1056b887ea65" containerID="9c6e66048455719a05e49fd6068d11524a5db88f7e8bbc27f4952110fcfee101" exitCode=137 Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.950052 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c49ded6e-3eb8-44f1-9125-5938485a644a","Type":"ContainerStarted","Data":"6cbf266ee9839a3fc3df341760e9daf7baedbae71fe18b5a3694b459743f58d6"} Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.951904 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"810023b7-c0cb-4c82-9d79-387562b0b0d1","Type":"ContainerStarted","Data":"e3a36877da691baf83b7b6d38093774147f48625cb890473b4f128573ba0ad24"} Dec 05 12:51:50 crc kubenswrapper[4809]: I1205 12:51:50.985122 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.096182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.261915 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.461893 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config\") pod \"73acb568-95c4-4d30-81bd-1056b887ea65\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.462530 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5992h\" (UniqueName: \"kubernetes.io/projected/73acb568-95c4-4d30-81bd-1056b887ea65-kube-api-access-5992h\") pod \"73acb568-95c4-4d30-81bd-1056b887ea65\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.462781 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config-secret\") pod \"73acb568-95c4-4d30-81bd-1056b887ea65\" (UID: \"73acb568-95c4-4d30-81bd-1056b887ea65\") " Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.471152 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73acb568-95c4-4d30-81bd-1056b887ea65-kube-api-access-5992h" (OuterVolumeSpecName: "kube-api-access-5992h") pod "73acb568-95c4-4d30-81bd-1056b887ea65" (UID: "73acb568-95c4-4d30-81bd-1056b887ea65"). InnerVolumeSpecName "kube-api-access-5992h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.500538 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "73acb568-95c4-4d30-81bd-1056b887ea65" (UID: "73acb568-95c4-4d30-81bd-1056b887ea65"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.537094 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "73acb568-95c4-4d30-81bd-1056b887ea65" (UID: "73acb568-95c4-4d30-81bd-1056b887ea65"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.564949 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.564989 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73acb568-95c4-4d30-81bd-1056b887ea65-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.565000 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5992h\" (UniqueName: \"kubernetes.io/projected/73acb568-95c4-4d30-81bd-1056b887ea65-kube-api-access-5992h\") on node \"crc\" DevicePath \"\"" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.722810 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.966525 4809 scope.go:117] "RemoveContainer" containerID="9c6e66048455719a05e49fd6068d11524a5db88f7e8bbc27f4952110fcfee101" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.967671 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.970732 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"374aeffe-f715-4f6e-af5b-4ff7f547d9c5","Type":"ContainerStarted","Data":"5f3704190ab9d7ed1b53d5af5bbe6203feb37459c4dc54fd36c7f10faf330ce7"} Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.977786 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c49ded6e-3eb8-44f1-9125-5938485a644a","Type":"ContainerStarted","Data":"a8bbb21dfd89dbcce8c62c57e9ed9610d664af66c956bda657c48a9334177397"} Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.981761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0381cceb-e34d-4a34-bfb0-8673a52237dc","Type":"ContainerStarted","Data":"82cda470f24d0aa633ec9d89a70db736829550b8696d1db0c3d2e2bcdfca4efb"} Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.991445 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"810023b7-c0cb-4c82-9d79-387562b0b0d1","Type":"ContainerStarted","Data":"9211474db0a18e49d8dd37040c7bcae40e92d6859f25e8e9bf466ee626af77b3"} Dec 05 12:51:51 crc kubenswrapper[4809]: I1205 12:51:51.991899 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 12:51:52 crc kubenswrapper[4809]: I1205 12:51:52.021481 4809 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="73acb568-95c4-4d30-81bd-1056b887ea65" podUID="c49ded6e-3eb8-44f1-9125-5938485a644a" Dec 05 12:51:52 crc kubenswrapper[4809]: I1205 12:51:52.023071 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.023056144 podStartE2EDuration="4.023056144s" podCreationTimestamp="2025-12-05 12:51:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:51:51.994761052 +0000 UTC m=+6207.385737630" watchObservedRunningTime="2025-12-05 12:51:52.023056144 +0000 UTC m=+6207.414032702" Dec 05 12:51:52 crc kubenswrapper[4809]: I1205 12:51:52.027148 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.398664285 podStartE2EDuration="4.027131483s" podCreationTimestamp="2025-12-05 12:51:48 +0000 UTC" firstStartedPulling="2025-12-05 12:51:50.234307617 +0000 UTC m=+6205.625284175" lastFinishedPulling="2025-12-05 12:51:50.862774815 +0000 UTC m=+6206.253751373" observedRunningTime="2025-12-05 12:51:52.018643405 +0000 UTC m=+6207.409619973" watchObservedRunningTime="2025-12-05 12:51:52.027131483 +0000 UTC m=+6207.418108041" Dec 05 12:51:52 crc kubenswrapper[4809]: I1205 12:51:52.884435 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73acb568-95c4-4d30-81bd-1056b887ea65" path="/var/lib/kubelet/pods/73acb568-95c4-4d30-81bd-1056b887ea65/volumes" Dec 05 12:51:57 crc kubenswrapper[4809]: I1205 12:51:57.044918 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"374aeffe-f715-4f6e-af5b-4ff7f547d9c5","Type":"ContainerStarted","Data":"83213f4d1c664a9a7dc01e68a039628c5f6f9ded8e79d099b75e0f8a3b61513f"} Dec 05 12:51:57 crc kubenswrapper[4809]: I1205 12:51:57.047526 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0381cceb-e34d-4a34-bfb0-8673a52237dc","Type":"ContainerStarted","Data":"8c3db9af7b2a45d545b7440209552f152be936e6238e55afe02374b3da09ff3d"} Dec 05 12:51:59 crc kubenswrapper[4809]: I1205 12:51:59.407224 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.157272 4809 generic.go:334] "Generic (PLEG): container finished" podID="374aeffe-f715-4f6e-af5b-4ff7f547d9c5" containerID="83213f4d1c664a9a7dc01e68a039628c5f6f9ded8e79d099b75e0f8a3b61513f" exitCode=0 Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.157359 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"374aeffe-f715-4f6e-af5b-4ff7f547d9c5","Type":"ContainerDied","Data":"83213f4d1c664a9a7dc01e68a039628c5f6f9ded8e79d099b75e0f8a3b61513f"} Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.301893 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lx79k"] Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.304122 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.317108 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx79k"] Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.415237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-utilities\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.415286 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nbj9\" (UniqueName: \"kubernetes.io/projected/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-kube-api-access-8nbj9\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.415316 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-catalog-content\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.519113 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-utilities\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.519173 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nbj9\" (UniqueName: \"kubernetes.io/projected/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-kube-api-access-8nbj9\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.519557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-catalog-content\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.519616 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-utilities\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.519916 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-catalog-content\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.538088 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nbj9\" (UniqueName: \"kubernetes.io/projected/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-kube-api-access-8nbj9\") pod \"redhat-marketplace-lx79k\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:05 crc kubenswrapper[4809]: I1205 12:52:05.637607 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:06 crc kubenswrapper[4809]: I1205 12:52:06.174256 4809 generic.go:334] "Generic (PLEG): container finished" podID="0381cceb-e34d-4a34-bfb0-8673a52237dc" containerID="8c3db9af7b2a45d545b7440209552f152be936e6238e55afe02374b3da09ff3d" exitCode=0 Dec 05 12:52:06 crc kubenswrapper[4809]: I1205 12:52:06.174430 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0381cceb-e34d-4a34-bfb0-8673a52237dc","Type":"ContainerDied","Data":"8c3db9af7b2a45d545b7440209552f152be936e6238e55afe02374b3da09ff3d"} Dec 05 12:52:06 crc kubenswrapper[4809]: I1205 12:52:06.243130 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx79k"] Dec 05 12:52:07 crc kubenswrapper[4809]: I1205 12:52:07.203195 4809 generic.go:334] "Generic (PLEG): container finished" podID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerID="a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04" exitCode=0 Dec 05 12:52:07 crc kubenswrapper[4809]: I1205 12:52:07.203406 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx79k" event={"ID":"2f2835fb-84aa-4316-9ea3-f136fc5a40d3","Type":"ContainerDied","Data":"a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04"} Dec 05 12:52:07 crc kubenswrapper[4809]: I1205 12:52:07.203435 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx79k" event={"ID":"2f2835fb-84aa-4316-9ea3-f136fc5a40d3","Type":"ContainerStarted","Data":"6e6f5bddaa0b4c2e7ba0f73367c08ffe439966d6a84140ea007bdcffacc1c5ae"} Dec 05 12:52:08 crc kubenswrapper[4809]: I1205 12:52:08.214425 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx79k" event={"ID":"2f2835fb-84aa-4316-9ea3-f136fc5a40d3","Type":"ContainerStarted","Data":"35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a"} Dec 05 12:52:08 crc kubenswrapper[4809]: I1205 12:52:08.216885 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"374aeffe-f715-4f6e-af5b-4ff7f547d9c5","Type":"ContainerStarted","Data":"f0efae1f5891517fc52420571f8ecf998fc0577afe06aaef60ce89d2c2e2131b"} Dec 05 12:52:09 crc kubenswrapper[4809]: I1205 12:52:09.227554 4809 generic.go:334] "Generic (PLEG): container finished" podID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerID="35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a" exitCode=0 Dec 05 12:52:09 crc kubenswrapper[4809]: I1205 12:52:09.227756 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx79k" event={"ID":"2f2835fb-84aa-4316-9ea3-f136fc5a40d3","Type":"ContainerDied","Data":"35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a"} Dec 05 12:52:11 crc kubenswrapper[4809]: I1205 12:52:11.247011 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"374aeffe-f715-4f6e-af5b-4ff7f547d9c5","Type":"ContainerStarted","Data":"db79ca723565918fa5c9e98816f7c9172fd8d615060aa1d6c7a47c2baa9a3d9b"} Dec 05 12:52:11 crc kubenswrapper[4809]: I1205 12:52:11.247991 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 05 12:52:11 crc kubenswrapper[4809]: I1205 12:52:11.249973 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 05 12:52:11 crc kubenswrapper[4809]: I1205 12:52:11.272616 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.821913318 podStartE2EDuration="22.272595363s" podCreationTimestamp="2025-12-05 12:51:49 +0000 UTC" firstStartedPulling="2025-12-05 12:51:51.004306683 +0000 UTC m=+6206.395283241" lastFinishedPulling="2025-12-05 12:52:07.454988728 +0000 UTC m=+6222.845965286" observedRunningTime="2025-12-05 12:52:11.267582898 +0000 UTC m=+6226.658559456" watchObservedRunningTime="2025-12-05 12:52:11.272595363 +0000 UTC m=+6226.663571921" Dec 05 12:52:13 crc kubenswrapper[4809]: I1205 12:52:13.302645 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx79k" event={"ID":"2f2835fb-84aa-4316-9ea3-f136fc5a40d3","Type":"ContainerStarted","Data":"4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88"} Dec 05 12:52:14 crc kubenswrapper[4809]: I1205 12:52:14.323056 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0381cceb-e34d-4a34-bfb0-8673a52237dc","Type":"ContainerStarted","Data":"864bd8cfed64c0dad2549c2a080b19e103f9301a1eda73b4e2d96657aa89804c"} Dec 05 12:52:15 crc kubenswrapper[4809]: I1205 12:52:15.638798 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:15 crc kubenswrapper[4809]: I1205 12:52:15.639274 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:15 crc kubenswrapper[4809]: I1205 12:52:15.723584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:15 crc kubenswrapper[4809]: I1205 12:52:15.753189 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lx79k" podStartSLOduration=7.572100026 podStartE2EDuration="10.753167064s" podCreationTimestamp="2025-12-05 12:52:05 +0000 UTC" firstStartedPulling="2025-12-05 12:52:07.372761996 +0000 UTC m=+6222.763738554" lastFinishedPulling="2025-12-05 12:52:10.553828994 +0000 UTC m=+6225.944805592" observedRunningTime="2025-12-05 12:52:13.322025384 +0000 UTC m=+6228.713001952" watchObservedRunningTime="2025-12-05 12:52:15.753167064 +0000 UTC m=+6231.144143652" Dec 05 12:52:18 crc kubenswrapper[4809]: I1205 12:52:18.365436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0381cceb-e34d-4a34-bfb0-8673a52237dc","Type":"ContainerStarted","Data":"4d156782b172d9085d27fc18099cd3f0de8c64e1c313e9a980ba5e3af6836409"} Dec 05 12:52:20 crc kubenswrapper[4809]: I1205 12:52:20.391837 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0381cceb-e34d-4a34-bfb0-8673a52237dc","Type":"ContainerStarted","Data":"4f0f8ff302c140444b5056e0dde41bdbe0f59319330abceb0be6c6c0c1cd1cb7"} Dec 05 12:52:20 crc kubenswrapper[4809]: I1205 12:52:20.429099 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.055320012 podStartE2EDuration="31.429082143s" podCreationTimestamp="2025-12-05 12:51:49 +0000 UTC" firstStartedPulling="2025-12-05 12:51:51.738669512 +0000 UTC m=+6207.129646070" lastFinishedPulling="2025-12-05 12:52:20.112431653 +0000 UTC m=+6235.503408201" observedRunningTime="2025-12-05 12:52:20.421613332 +0000 UTC m=+6235.812589900" watchObservedRunningTime="2025-12-05 12:52:20.429082143 +0000 UTC m=+6235.820058701" Dec 05 12:52:21 crc kubenswrapper[4809]: I1205 12:52:21.096915 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 05 12:52:21 crc kubenswrapper[4809]: I1205 12:52:21.097282 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 05 12:52:21 crc kubenswrapper[4809]: I1205 12:52:21.099981 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 05 12:52:21 crc kubenswrapper[4809]: I1205 12:52:21.426733 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 05 12:52:24 crc kubenswrapper[4809]: I1205 12:52:24.062285 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wmrmc"] Dec 05 12:52:24 crc kubenswrapper[4809]: I1205 12:52:24.080915 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wmrmc"] Dec 05 12:52:24 crc kubenswrapper[4809]: I1205 12:52:24.887844 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dccc2924-a397-4c16-961b-26a510522a1c" path="/var/lib/kubelet/pods/dccc2924-a397-4c16-961b-26a510522a1c/volumes" Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.043523 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bdsxj"] Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.059055 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lnxhz"] Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.068306 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bdsxj"] Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.077088 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-03e6-account-create-update-8vfsh"] Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.086711 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lnxhz"] Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.097834 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-03e6-account-create-update-8vfsh"] Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.734183 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:25 crc kubenswrapper[4809]: I1205 12:52:25.827680 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx79k"] Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.046934 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9e4c-account-create-update-wl2w2"] Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.056422 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9e4c-account-create-update-wl2w2"] Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.064806 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-abbc-account-create-update-dql8l"] Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.074239 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-abbc-account-create-update-dql8l"] Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.487005 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lx79k" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="registry-server" containerID="cri-o://4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88" gracePeriod=2 Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.884110 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="103c4a83-b894-416e-ab74-56ee1f206e44" path="/var/lib/kubelet/pods/103c4a83-b894-416e-ab74-56ee1f206e44/volumes" Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.885199 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="443f06d4-d0db-4c99-8497-ffa1ffea10a8" path="/var/lib/kubelet/pods/443f06d4-d0db-4c99-8497-ffa1ffea10a8/volumes" Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.885819 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d624929-bbe0-4674-8aa5-6415feaacb60" path="/var/lib/kubelet/pods/4d624929-bbe0-4674-8aa5-6415feaacb60/volumes" Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.886402 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d18b96-a42f-47fe-abcc-6a6ec909114e" path="/var/lib/kubelet/pods/e2d18b96-a42f-47fe-abcc-6a6ec909114e/volumes" Dec 05 12:52:26 crc kubenswrapper[4809]: I1205 12:52:26.887995 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5074f64-b9a7-4675-b323-1a3af9beeb7f" path="/var/lib/kubelet/pods/e5074f64-b9a7-4675-b323-1a3af9beeb7f/volumes" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.025122 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.030978 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-catalog-content\") pod \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.031039 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-utilities\") pod \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.031076 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nbj9\" (UniqueName: \"kubernetes.io/projected/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-kube-api-access-8nbj9\") pod \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\" (UID: \"2f2835fb-84aa-4316-9ea3-f136fc5a40d3\") " Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.031866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-utilities" (OuterVolumeSpecName: "utilities") pod "2f2835fb-84aa-4316-9ea3-f136fc5a40d3" (UID: "2f2835fb-84aa-4316-9ea3-f136fc5a40d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.036842 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-kube-api-access-8nbj9" (OuterVolumeSpecName: "kube-api-access-8nbj9") pod "2f2835fb-84aa-4316-9ea3-f136fc5a40d3" (UID: "2f2835fb-84aa-4316-9ea3-f136fc5a40d3"). InnerVolumeSpecName "kube-api-access-8nbj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.054942 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f2835fb-84aa-4316-9ea3-f136fc5a40d3" (UID: "2f2835fb-84aa-4316-9ea3-f136fc5a40d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.132968 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.133002 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.133013 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nbj9\" (UniqueName: \"kubernetes.io/projected/2f2835fb-84aa-4316-9ea3-f136fc5a40d3-kube-api-access-8nbj9\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.499558 4809 generic.go:334] "Generic (PLEG): container finished" podID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerID="4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88" exitCode=0 Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.499610 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx79k" event={"ID":"2f2835fb-84aa-4316-9ea3-f136fc5a40d3","Type":"ContainerDied","Data":"4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88"} Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.499650 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lx79k" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.499677 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lx79k" event={"ID":"2f2835fb-84aa-4316-9ea3-f136fc5a40d3","Type":"ContainerDied","Data":"6e6f5bddaa0b4c2e7ba0f73367c08ffe439966d6a84140ea007bdcffacc1c5ae"} Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.499704 4809 scope.go:117] "RemoveContainer" containerID="4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.527193 4809 scope.go:117] "RemoveContainer" containerID="35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.543175 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:52:27 crc kubenswrapper[4809]: E1205 12:52:27.543777 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="extract-content" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.543801 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="extract-content" Dec 05 12:52:27 crc kubenswrapper[4809]: E1205 12:52:27.543824 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="extract-utilities" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.543833 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="extract-utilities" Dec 05 12:52:27 crc kubenswrapper[4809]: E1205 12:52:27.543850 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="registry-server" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.543859 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="registry-server" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.544122 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" containerName="registry-server" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.547188 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.552721 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.553141 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.559694 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx79k"] Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.577553 4809 scope.go:117] "RemoveContainer" containerID="a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.577722 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lx79k"] Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.583172 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.633600 4809 scope.go:117] "RemoveContainer" containerID="4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88" Dec 05 12:52:27 crc kubenswrapper[4809]: E1205 12:52:27.634300 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88\": container with ID starting with 4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88 not found: ID does not exist" containerID="4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.634339 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88"} err="failed to get container status \"4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88\": rpc error: code = NotFound desc = could not find container \"4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88\": container with ID starting with 4d690f76000508e8e4a33821e8774e8e51a30cc1290b7a659067f5aa31e6aa88 not found: ID does not exist" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.634364 4809 scope.go:117] "RemoveContainer" containerID="35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a" Dec 05 12:52:27 crc kubenswrapper[4809]: E1205 12:52:27.634591 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a\": container with ID starting with 35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a not found: ID does not exist" containerID="35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.634618 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a"} err="failed to get container status \"35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a\": rpc error: code = NotFound desc = could not find container \"35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a\": container with ID starting with 35579ecf5e9405aae939202f17527ae513d87c3273fa52c9547def35871a009a not found: ID does not exist" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.634718 4809 scope.go:117] "RemoveContainer" containerID="a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04" Dec 05 12:52:27 crc kubenswrapper[4809]: E1205 12:52:27.634935 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04\": container with ID starting with a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04 not found: ID does not exist" containerID="a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.634960 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04"} err="failed to get container status \"a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04\": rpc error: code = NotFound desc = could not find container \"a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04\": container with ID starting with a7ef311100b3da58f7a94f82bce15fdb5b72395731c5679b14656da1fa32bb04 not found: ID does not exist" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.645290 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.645354 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.645592 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-run-httpd\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.645684 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx52p\" (UniqueName: \"kubernetes.io/projected/034fe145-c83b-4b03-b666-d667c3d242a6-kube-api-access-qx52p\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.645840 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-log-httpd\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.645875 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-scripts\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.645999 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-config-data\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.748803 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.748844 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.748902 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-run-httpd\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.748931 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx52p\" (UniqueName: \"kubernetes.io/projected/034fe145-c83b-4b03-b666-d667c3d242a6-kube-api-access-qx52p\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.748983 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-log-httpd\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.749001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-scripts\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.749049 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-config-data\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.754519 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-config-data\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.756562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-run-httpd\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.756667 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-log-httpd\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.764318 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.764423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.764600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-scripts\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.774992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx52p\" (UniqueName: \"kubernetes.io/projected/034fe145-c83b-4b03-b666-d667c3d242a6-kube-api-access-qx52p\") pod \"ceilometer-0\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " pod="openstack/ceilometer-0" Dec 05 12:52:27 crc kubenswrapper[4809]: I1205 12:52:27.954567 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:52:28 crc kubenswrapper[4809]: I1205 12:52:28.522159 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:52:28 crc kubenswrapper[4809]: I1205 12:52:28.883357 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f2835fb-84aa-4316-9ea3-f136fc5a40d3" path="/var/lib/kubelet/pods/2f2835fb-84aa-4316-9ea3-f136fc5a40d3/volumes" Dec 05 12:52:29 crc kubenswrapper[4809]: I1205 12:52:29.522844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerStarted","Data":"3f856f8295dd1000ef901ad36705df999653479b2c3afee565a43673c15ac28c"} Dec 05 12:52:29 crc kubenswrapper[4809]: I1205 12:52:29.523174 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerStarted","Data":"c9e1b13a5e4e0dd728d30888c57d9e410f5a7ac5b75651bb0fba768fd7bae755"} Dec 05 12:52:30 crc kubenswrapper[4809]: I1205 12:52:30.536721 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerStarted","Data":"0c9712a6605652ac0fe2f7207befe9f417e070a32becd9d5062c1f7e86319442"} Dec 05 12:52:30 crc kubenswrapper[4809]: I1205 12:52:30.538006 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerStarted","Data":"2b9b5d181eb96c8a949dbd732200fc47ea41f012754125d0d40bafed33b9adee"} Dec 05 12:52:33 crc kubenswrapper[4809]: I1205 12:52:33.590534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerStarted","Data":"b92a0e46f69b0ddf69d1f27dd0e020a6bff979a620ae5e67cb7bddaab66588da"} Dec 05 12:52:33 crc kubenswrapper[4809]: I1205 12:52:33.591087 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 12:52:33 crc kubenswrapper[4809]: I1205 12:52:33.613590 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.405998309 podStartE2EDuration="6.613571216s" podCreationTimestamp="2025-12-05 12:52:27 +0000 UTC" firstStartedPulling="2025-12-05 12:52:28.517220127 +0000 UTC m=+6243.908196695" lastFinishedPulling="2025-12-05 12:52:32.724793044 +0000 UTC m=+6248.115769602" observedRunningTime="2025-12-05 12:52:33.612550288 +0000 UTC m=+6249.003526856" watchObservedRunningTime="2025-12-05 12:52:33.613571216 +0000 UTC m=+6249.004547794" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.181594 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-pkhwb"] Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.184684 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.198220 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-pkhwb"] Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.281757 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-8233-account-create-update-rwtqt"] Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.283139 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.288064 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.321745 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-8233-account-create-update-rwtqt"] Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.323225 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-operator-scripts\") pod \"aodh-db-create-pkhwb\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.323260 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndsfz\" (UniqueName: \"kubernetes.io/projected/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-kube-api-access-ndsfz\") pod \"aodh-db-create-pkhwb\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.425414 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563aa06-65da-4383-ab21-39c19d9ab0c3-operator-scripts\") pod \"aodh-8233-account-create-update-rwtqt\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.426117 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6dxk\" (UniqueName: \"kubernetes.io/projected/3563aa06-65da-4383-ab21-39c19d9ab0c3-kube-api-access-m6dxk\") pod \"aodh-8233-account-create-update-rwtqt\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.426416 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-operator-scripts\") pod \"aodh-db-create-pkhwb\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.426464 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndsfz\" (UniqueName: \"kubernetes.io/projected/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-kube-api-access-ndsfz\") pod \"aodh-db-create-pkhwb\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.427356 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-operator-scripts\") pod \"aodh-db-create-pkhwb\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.445796 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndsfz\" (UniqueName: \"kubernetes.io/projected/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-kube-api-access-ndsfz\") pod \"aodh-db-create-pkhwb\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.509463 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.528846 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563aa06-65da-4383-ab21-39c19d9ab0c3-operator-scripts\") pod \"aodh-8233-account-create-update-rwtqt\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.528936 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6dxk\" (UniqueName: \"kubernetes.io/projected/3563aa06-65da-4383-ab21-39c19d9ab0c3-kube-api-access-m6dxk\") pod \"aodh-8233-account-create-update-rwtqt\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.529805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563aa06-65da-4383-ab21-39c19d9ab0c3-operator-scripts\") pod \"aodh-8233-account-create-update-rwtqt\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.546445 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6dxk\" (UniqueName: \"kubernetes.io/projected/3563aa06-65da-4383-ab21-39c19d9ab0c3-kube-api-access-m6dxk\") pod \"aodh-8233-account-create-update-rwtqt\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:39 crc kubenswrapper[4809]: I1205 12:52:39.605523 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.039140 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dpbn8"] Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.056343 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dpbn8"] Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.065474 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-pkhwb"] Dec 05 12:52:40 crc kubenswrapper[4809]: W1205 12:52:40.212041 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3563aa06_65da_4383_ab21_39c19d9ab0c3.slice/crio-e29e242e4b70be236ea9c35237856bb2057633f74e3a99e9e210cfd3c8dc3646 WatchSource:0}: Error finding container e29e242e4b70be236ea9c35237856bb2057633f74e3a99e9e210cfd3c8dc3646: Status 404 returned error can't find the container with id e29e242e4b70be236ea9c35237856bb2057633f74e3a99e9e210cfd3c8dc3646 Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.216146 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-8233-account-create-update-rwtqt"] Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.698139 4809 generic.go:334] "Generic (PLEG): container finished" podID="1f2ff1bb-c9f5-4772-814c-71ed1d78ae43" containerID="8d5c13451a3cd9cda653ba05d893de8041bb378443aae68e46fc8ae76fd82eb9" exitCode=0 Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.698234 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pkhwb" event={"ID":"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43","Type":"ContainerDied","Data":"8d5c13451a3cd9cda653ba05d893de8041bb378443aae68e46fc8ae76fd82eb9"} Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.698268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pkhwb" event={"ID":"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43","Type":"ContainerStarted","Data":"7f01c4b0303a033f31c0ef2d415462baa0ad3b94cc9a280b52698101029cdb8f"} Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.699734 4809 generic.go:334] "Generic (PLEG): container finished" podID="3563aa06-65da-4383-ab21-39c19d9ab0c3" containerID="64ed934b366e0c56e6de906ae92943b3f634dc94089bb7eaad5c0940d8603ab3" exitCode=0 Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.699769 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8233-account-create-update-rwtqt" event={"ID":"3563aa06-65da-4383-ab21-39c19d9ab0c3","Type":"ContainerDied","Data":"64ed934b366e0c56e6de906ae92943b3f634dc94089bb7eaad5c0940d8603ab3"} Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.699789 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8233-account-create-update-rwtqt" event={"ID":"3563aa06-65da-4383-ab21-39c19d9ab0c3","Type":"ContainerStarted","Data":"e29e242e4b70be236ea9c35237856bb2057633f74e3a99e9e210cfd3c8dc3646"} Dec 05 12:52:40 crc kubenswrapper[4809]: I1205 12:52:40.892227 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30302f31-4bc7-4fce-b7c5-61ad600ade36" path="/var/lib/kubelet/pods/30302f31-4bc7-4fce-b7c5-61ad600ade36/volumes" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.096845 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.186844 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563aa06-65da-4383-ab21-39c19d9ab0c3-operator-scripts\") pod \"3563aa06-65da-4383-ab21-39c19d9ab0c3\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.186979 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6dxk\" (UniqueName: \"kubernetes.io/projected/3563aa06-65da-4383-ab21-39c19d9ab0c3-kube-api-access-m6dxk\") pod \"3563aa06-65da-4383-ab21-39c19d9ab0c3\" (UID: \"3563aa06-65da-4383-ab21-39c19d9ab0c3\") " Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.187312 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3563aa06-65da-4383-ab21-39c19d9ab0c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3563aa06-65da-4383-ab21-39c19d9ab0c3" (UID: "3563aa06-65da-4383-ab21-39c19d9ab0c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.187854 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3563aa06-65da-4383-ab21-39c19d9ab0c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.192022 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3563aa06-65da-4383-ab21-39c19d9ab0c3-kube-api-access-m6dxk" (OuterVolumeSpecName: "kube-api-access-m6dxk") pod "3563aa06-65da-4383-ab21-39c19d9ab0c3" (UID: "3563aa06-65da-4383-ab21-39c19d9ab0c3"). InnerVolumeSpecName "kube-api-access-m6dxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.242506 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.289919 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6dxk\" (UniqueName: \"kubernetes.io/projected/3563aa06-65da-4383-ab21-39c19d9ab0c3-kube-api-access-m6dxk\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.391660 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-operator-scripts\") pod \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.391714 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndsfz\" (UniqueName: \"kubernetes.io/projected/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-kube-api-access-ndsfz\") pod \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\" (UID: \"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43\") " Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.392197 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1f2ff1bb-c9f5-4772-814c-71ed1d78ae43" (UID: "1f2ff1bb-c9f5-4772-814c-71ed1d78ae43"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.395062 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-kube-api-access-ndsfz" (OuterVolumeSpecName: "kube-api-access-ndsfz") pod "1f2ff1bb-c9f5-4772-814c-71ed1d78ae43" (UID: "1f2ff1bb-c9f5-4772-814c-71ed1d78ae43"). InnerVolumeSpecName "kube-api-access-ndsfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.494111 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.494154 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndsfz\" (UniqueName: \"kubernetes.io/projected/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43-kube-api-access-ndsfz\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.727924 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-8233-account-create-update-rwtqt" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.727822 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-8233-account-create-update-rwtqt" event={"ID":"3563aa06-65da-4383-ab21-39c19d9ab0c3","Type":"ContainerDied","Data":"e29e242e4b70be236ea9c35237856bb2057633f74e3a99e9e210cfd3c8dc3646"} Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.728200 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e29e242e4b70be236ea9c35237856bb2057633f74e3a99e9e210cfd3c8dc3646" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.729761 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pkhwb" event={"ID":"1f2ff1bb-c9f5-4772-814c-71ed1d78ae43","Type":"ContainerDied","Data":"7f01c4b0303a033f31c0ef2d415462baa0ad3b94cc9a280b52698101029cdb8f"} Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.729813 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f01c4b0303a033f31c0ef2d415462baa0ad3b94cc9a280b52698101029cdb8f" Dec 05 12:52:42 crc kubenswrapper[4809]: I1205 12:52:42.729819 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkhwb" Dec 05 12:52:43 crc kubenswrapper[4809]: I1205 12:52:43.445083 4809 scope.go:117] "RemoveContainer" containerID="bcd812d0b2fdabc8e3b5af5ad8152747a2b26ce6247f8e4bd6507555f973fdba" Dec 05 12:52:43 crc kubenswrapper[4809]: I1205 12:52:43.492548 4809 scope.go:117] "RemoveContainer" containerID="b5b53c981c27e44c124008d1ef0cad30dfac017d80a079817d0f62c4f2e3247b" Dec 05 12:52:43 crc kubenswrapper[4809]: I1205 12:52:43.547558 4809 scope.go:117] "RemoveContainer" containerID="d3806d8e20c39f4ad1e496494af6b9d4ebd21cf1eab899f903cd4ef83b63be7b" Dec 05 12:52:43 crc kubenswrapper[4809]: I1205 12:52:43.601864 4809 scope.go:117] "RemoveContainer" containerID="4f45e53c3e7b3f69dd1517f4f5e0b54123f4c25fcd1cc38c842e31da97f6f3e2" Dec 05 12:52:43 crc kubenswrapper[4809]: I1205 12:52:43.660592 4809 scope.go:117] "RemoveContainer" containerID="b0f8d2ae810086c971d66875a1ccf6202fde508960c503652f49a7d62fb56e69" Dec 05 12:52:43 crc kubenswrapper[4809]: I1205 12:52:43.727932 4809 scope.go:117] "RemoveContainer" containerID="0456a8fda8ba44a040478dfa2f1796ce0e518eeb81de9e3bc1a22c6066761a23" Dec 05 12:52:43 crc kubenswrapper[4809]: I1205 12:52:43.751306 4809 scope.go:117] "RemoveContainer" containerID="589f7452101fa440c93234a048b4e6cccdd2dd1b818482625b043d71467c6011" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.657961 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-5sxdg"] Dec 05 12:52:44 crc kubenswrapper[4809]: E1205 12:52:44.658461 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f2ff1bb-c9f5-4772-814c-71ed1d78ae43" containerName="mariadb-database-create" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.658500 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f2ff1bb-c9f5-4772-814c-71ed1d78ae43" containerName="mariadb-database-create" Dec 05 12:52:44 crc kubenswrapper[4809]: E1205 12:52:44.658513 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3563aa06-65da-4383-ab21-39c19d9ab0c3" containerName="mariadb-account-create-update" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.658522 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3563aa06-65da-4383-ab21-39c19d9ab0c3" containerName="mariadb-account-create-update" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.658833 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3563aa06-65da-4383-ab21-39c19d9ab0c3" containerName="mariadb-account-create-update" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.658853 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f2ff1bb-c9f5-4772-814c-71ed1d78ae43" containerName="mariadb-database-create" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.659700 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.664660 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-5sxdg"] Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.708397 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.708659 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dqmf6" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.708868 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.708921 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.743447 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-scripts\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.743537 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-combined-ca-bundle\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.743612 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-config-data\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.743691 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znl5q\" (UniqueName: \"kubernetes.io/projected/b6018b72-920a-4d96-8351-45e1ee9810f6-kube-api-access-znl5q\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.845707 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-scripts\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.845842 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-combined-ca-bundle\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.845951 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-config-data\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.846026 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znl5q\" (UniqueName: \"kubernetes.io/projected/b6018b72-920a-4d96-8351-45e1ee9810f6-kube-api-access-znl5q\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.851600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-scripts\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.852646 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-combined-ca-bundle\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.853113 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-config-data\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:44 crc kubenswrapper[4809]: I1205 12:52:44.863510 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znl5q\" (UniqueName: \"kubernetes.io/projected/b6018b72-920a-4d96-8351-45e1ee9810f6-kube-api-access-znl5q\") pod \"aodh-db-sync-5sxdg\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:45 crc kubenswrapper[4809]: I1205 12:52:45.046173 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:45 crc kubenswrapper[4809]: W1205 12:52:45.537853 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6018b72_920a_4d96_8351_45e1ee9810f6.slice/crio-0a8d2e2adfc9661d02b1d5f9404b2e2ec4ba911b537ed180bb4cd5dd37397401 WatchSource:0}: Error finding container 0a8d2e2adfc9661d02b1d5f9404b2e2ec4ba911b537ed180bb4cd5dd37397401: Status 404 returned error can't find the container with id 0a8d2e2adfc9661d02b1d5f9404b2e2ec4ba911b537ed180bb4cd5dd37397401 Dec 05 12:52:45 crc kubenswrapper[4809]: I1205 12:52:45.538840 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-5sxdg"] Dec 05 12:52:45 crc kubenswrapper[4809]: I1205 12:52:45.788707 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5sxdg" event={"ID":"b6018b72-920a-4d96-8351-45e1ee9810f6","Type":"ContainerStarted","Data":"0a8d2e2adfc9661d02b1d5f9404b2e2ec4ba911b537ed180bb4cd5dd37397401"} Dec 05 12:52:50 crc kubenswrapper[4809]: I1205 12:52:50.851012 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5sxdg" event={"ID":"b6018b72-920a-4d96-8351-45e1ee9810f6","Type":"ContainerStarted","Data":"cde4ef13a0db116a2b4cd335147e623b31b5889a518bf06b98f251f29002d7f4"} Dec 05 12:52:50 crc kubenswrapper[4809]: I1205 12:52:50.877567 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-5sxdg" podStartSLOduration=2.673344647 podStartE2EDuration="6.877553523s" podCreationTimestamp="2025-12-05 12:52:44 +0000 UTC" firstStartedPulling="2025-12-05 12:52:45.539508 +0000 UTC m=+6260.930484568" lastFinishedPulling="2025-12-05 12:52:49.743716886 +0000 UTC m=+6265.134693444" observedRunningTime="2025-12-05 12:52:50.867615236 +0000 UTC m=+6266.258591804" watchObservedRunningTime="2025-12-05 12:52:50.877553523 +0000 UTC m=+6266.268530081" Dec 05 12:52:52 crc kubenswrapper[4809]: I1205 12:52:52.870119 4809 generic.go:334] "Generic (PLEG): container finished" podID="b6018b72-920a-4d96-8351-45e1ee9810f6" containerID="cde4ef13a0db116a2b4cd335147e623b31b5889a518bf06b98f251f29002d7f4" exitCode=0 Dec 05 12:52:52 crc kubenswrapper[4809]: I1205 12:52:52.870215 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5sxdg" event={"ID":"b6018b72-920a-4d96-8351-45e1ee9810f6","Type":"ContainerDied","Data":"cde4ef13a0db116a2b4cd335147e623b31b5889a518bf06b98f251f29002d7f4"} Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.274948 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.346908 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-scripts\") pod \"b6018b72-920a-4d96-8351-45e1ee9810f6\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.347092 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-config-data\") pod \"b6018b72-920a-4d96-8351-45e1ee9810f6\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.347211 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znl5q\" (UniqueName: \"kubernetes.io/projected/b6018b72-920a-4d96-8351-45e1ee9810f6-kube-api-access-znl5q\") pod \"b6018b72-920a-4d96-8351-45e1ee9810f6\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.347239 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-combined-ca-bundle\") pod \"b6018b72-920a-4d96-8351-45e1ee9810f6\" (UID: \"b6018b72-920a-4d96-8351-45e1ee9810f6\") " Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.354402 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6018b72-920a-4d96-8351-45e1ee9810f6-kube-api-access-znl5q" (OuterVolumeSpecName: "kube-api-access-znl5q") pod "b6018b72-920a-4d96-8351-45e1ee9810f6" (UID: "b6018b72-920a-4d96-8351-45e1ee9810f6"). InnerVolumeSpecName "kube-api-access-znl5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.356233 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-scripts" (OuterVolumeSpecName: "scripts") pod "b6018b72-920a-4d96-8351-45e1ee9810f6" (UID: "b6018b72-920a-4d96-8351-45e1ee9810f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.381810 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-config-data" (OuterVolumeSpecName: "config-data") pod "b6018b72-920a-4d96-8351-45e1ee9810f6" (UID: "b6018b72-920a-4d96-8351-45e1ee9810f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.388621 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6018b72-920a-4d96-8351-45e1ee9810f6" (UID: "b6018b72-920a-4d96-8351-45e1ee9810f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.449580 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.449646 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znl5q\" (UniqueName: \"kubernetes.io/projected/b6018b72-920a-4d96-8351-45e1ee9810f6-kube-api-access-znl5q\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.449659 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.449671 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6018b72-920a-4d96-8351-45e1ee9810f6-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.924964 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5sxdg" event={"ID":"b6018b72-920a-4d96-8351-45e1ee9810f6","Type":"ContainerDied","Data":"0a8d2e2adfc9661d02b1d5f9404b2e2ec4ba911b537ed180bb4cd5dd37397401"} Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.927741 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a8d2e2adfc9661d02b1d5f9404b2e2ec4ba911b537ed180bb4cd5dd37397401" Dec 05 12:52:54 crc kubenswrapper[4809]: I1205 12:52:54.927955 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5sxdg" Dec 05 12:52:57 crc kubenswrapper[4809]: I1205 12:52:57.963718 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.052386 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b6jfm"] Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.072425 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-snglg"] Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.083649 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-snglg"] Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.094290 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b6jfm"] Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.380229 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 05 12:52:59 crc kubenswrapper[4809]: E1205 12:52:59.381016 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6018b72-920a-4d96-8351-45e1ee9810f6" containerName="aodh-db-sync" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.381127 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6018b72-920a-4d96-8351-45e1ee9810f6" containerName="aodh-db-sync" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.381505 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6018b72-920a-4d96-8351-45e1ee9810f6" containerName="aodh-db-sync" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.383567 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.385884 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.387183 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dqmf6" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.387384 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.388877 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.464090 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-scripts\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.464219 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwz5t\" (UniqueName: \"kubernetes.io/projected/9768daa9-0e4c-4f29-bb4d-0c548a818691-kube-api-access-bwz5t\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.464429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-config-data\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.464482 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-combined-ca-bundle\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.566966 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-scripts\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.567096 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwz5t\" (UniqueName: \"kubernetes.io/projected/9768daa9-0e4c-4f29-bb4d-0c548a818691-kube-api-access-bwz5t\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.567179 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-config-data\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.567197 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-combined-ca-bundle\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.574065 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-config-data\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.576996 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-scripts\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.591252 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwz5t\" (UniqueName: \"kubernetes.io/projected/9768daa9-0e4c-4f29-bb4d-0c548a818691-kube-api-access-bwz5t\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.591304 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9768daa9-0e4c-4f29-bb4d-0c548a818691-combined-ca-bundle\") pod \"aodh-0\" (UID: \"9768daa9-0e4c-4f29-bb4d-0c548a818691\") " pod="openstack/aodh-0" Dec 05 12:52:59 crc kubenswrapper[4809]: I1205 12:52:59.707488 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 12:53:00 crc kubenswrapper[4809]: I1205 12:53:00.303904 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 12:53:00 crc kubenswrapper[4809]: I1205 12:53:00.888048 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47917734-1cb9-4e2b-b016-36d5a57aa79c" path="/var/lib/kubelet/pods/47917734-1cb9-4e2b-b016-36d5a57aa79c/volumes" Dec 05 12:53:00 crc kubenswrapper[4809]: I1205 12:53:00.889083 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb2a36fb-5edb-4b55-bccf-df95e3d73500" path="/var/lib/kubelet/pods/fb2a36fb-5edb-4b55-bccf-df95e3d73500/volumes" Dec 05 12:53:00 crc kubenswrapper[4809]: I1205 12:53:00.998168 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9768daa9-0e4c-4f29-bb4d-0c548a818691","Type":"ContainerStarted","Data":"de33dc21be93d4d36dcbba710c20fd406eda4d220fd71d82255828ed3ee98b3b"} Dec 05 12:53:01 crc kubenswrapper[4809]: I1205 12:53:01.325556 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:01 crc kubenswrapper[4809]: I1205 12:53:01.326065 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-central-agent" containerID="cri-o://3f856f8295dd1000ef901ad36705df999653479b2c3afee565a43673c15ac28c" gracePeriod=30 Dec 05 12:53:01 crc kubenswrapper[4809]: I1205 12:53:01.326236 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="proxy-httpd" containerID="cri-o://b92a0e46f69b0ddf69d1f27dd0e020a6bff979a620ae5e67cb7bddaab66588da" gracePeriod=30 Dec 05 12:53:01 crc kubenswrapper[4809]: I1205 12:53:01.326274 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="sg-core" containerID="cri-o://0c9712a6605652ac0fe2f7207befe9f417e070a32becd9d5062c1f7e86319442" gracePeriod=30 Dec 05 12:53:01 crc kubenswrapper[4809]: I1205 12:53:01.326309 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-notification-agent" containerID="cri-o://2b9b5d181eb96c8a949dbd732200fc47ea41f012754125d0d40bafed33b9adee" gracePeriod=30 Dec 05 12:53:02 crc kubenswrapper[4809]: I1205 12:53:02.016876 4809 generic.go:334] "Generic (PLEG): container finished" podID="034fe145-c83b-4b03-b666-d667c3d242a6" containerID="b92a0e46f69b0ddf69d1f27dd0e020a6bff979a620ae5e67cb7bddaab66588da" exitCode=0 Dec 05 12:53:02 crc kubenswrapper[4809]: I1205 12:53:02.016950 4809 generic.go:334] "Generic (PLEG): container finished" podID="034fe145-c83b-4b03-b666-d667c3d242a6" containerID="0c9712a6605652ac0fe2f7207befe9f417e070a32becd9d5062c1f7e86319442" exitCode=2 Dec 05 12:53:02 crc kubenswrapper[4809]: I1205 12:53:02.017040 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerDied","Data":"b92a0e46f69b0ddf69d1f27dd0e020a6bff979a620ae5e67cb7bddaab66588da"} Dec 05 12:53:02 crc kubenswrapper[4809]: I1205 12:53:02.017122 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerDied","Data":"0c9712a6605652ac0fe2f7207befe9f417e070a32becd9d5062c1f7e86319442"} Dec 05 12:53:02 crc kubenswrapper[4809]: I1205 12:53:02.036201 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9768daa9-0e4c-4f29-bb4d-0c548a818691","Type":"ContainerStarted","Data":"520c28e1294ed210e21191f3b4f1bf6f7b2ace7018465c204450b2a1c1c982d5"} Dec 05 12:53:03 crc kubenswrapper[4809]: I1205 12:53:03.052413 4809 generic.go:334] "Generic (PLEG): container finished" podID="034fe145-c83b-4b03-b666-d667c3d242a6" containerID="3f856f8295dd1000ef901ad36705df999653479b2c3afee565a43673c15ac28c" exitCode=0 Dec 05 12:53:03 crc kubenswrapper[4809]: I1205 12:53:03.052512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerDied","Data":"3f856f8295dd1000ef901ad36705df999653479b2c3afee565a43673c15ac28c"} Dec 05 12:53:03 crc kubenswrapper[4809]: I1205 12:53:03.055233 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9768daa9-0e4c-4f29-bb4d-0c548a818691","Type":"ContainerStarted","Data":"11db839c1733c8ec72452e07096c5a5a28ec011b118aa40409f4543363a27d4f"} Dec 05 12:53:06 crc kubenswrapper[4809]: I1205 12:53:06.102235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9768daa9-0e4c-4f29-bb4d-0c548a818691","Type":"ContainerStarted","Data":"a518168fa774178fc7c7c0534d757eea313a83cbc5319cc5cd16cb92f46cd27d"} Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.150785 4809 generic.go:334] "Generic (PLEG): container finished" podID="034fe145-c83b-4b03-b666-d667c3d242a6" containerID="2b9b5d181eb96c8a949dbd732200fc47ea41f012754125d0d40bafed33b9adee" exitCode=0 Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.150953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerDied","Data":"2b9b5d181eb96c8a949dbd732200fc47ea41f012754125d0d40bafed33b9adee"} Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.151286 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"034fe145-c83b-4b03-b666-d667c3d242a6","Type":"ContainerDied","Data":"c9e1b13a5e4e0dd728d30888c57d9e410f5a7ac5b75651bb0fba768fd7bae755"} Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.151304 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9e1b13a5e4e0dd728d30888c57d9e410f5a7ac5b75651bb0fba768fd7bae755" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.154776 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9768daa9-0e4c-4f29-bb4d-0c548a818691","Type":"ContainerStarted","Data":"356876552cf5922e56611a54718dd962523274a61593c3fda4c428f0d602d077"} Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.181407 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.795989734 podStartE2EDuration="8.181374784s" podCreationTimestamp="2025-12-05 12:52:59 +0000 UTC" firstStartedPulling="2025-12-05 12:53:00.313388229 +0000 UTC m=+6275.704364787" lastFinishedPulling="2025-12-05 12:53:06.698773279 +0000 UTC m=+6282.089749837" observedRunningTime="2025-12-05 12:53:07.179461403 +0000 UTC m=+6282.570437961" watchObservedRunningTime="2025-12-05 12:53:07.181374784 +0000 UTC m=+6282.572351342" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.196425 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.268461 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-config-data\") pod \"034fe145-c83b-4b03-b666-d667c3d242a6\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.268543 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-run-httpd\") pod \"034fe145-c83b-4b03-b666-d667c3d242a6\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.268615 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-scripts\") pod \"034fe145-c83b-4b03-b666-d667c3d242a6\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.268652 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-log-httpd\") pod \"034fe145-c83b-4b03-b666-d667c3d242a6\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.268725 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx52p\" (UniqueName: \"kubernetes.io/projected/034fe145-c83b-4b03-b666-d667c3d242a6-kube-api-access-qx52p\") pod \"034fe145-c83b-4b03-b666-d667c3d242a6\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.268780 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-combined-ca-bundle\") pod \"034fe145-c83b-4b03-b666-d667c3d242a6\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.268831 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-sg-core-conf-yaml\") pod \"034fe145-c83b-4b03-b666-d667c3d242a6\" (UID: \"034fe145-c83b-4b03-b666-d667c3d242a6\") " Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.270461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "034fe145-c83b-4b03-b666-d667c3d242a6" (UID: "034fe145-c83b-4b03-b666-d667c3d242a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.270791 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "034fe145-c83b-4b03-b666-d667c3d242a6" (UID: "034fe145-c83b-4b03-b666-d667c3d242a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.273856 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034fe145-c83b-4b03-b666-d667c3d242a6-kube-api-access-qx52p" (OuterVolumeSpecName: "kube-api-access-qx52p") pod "034fe145-c83b-4b03-b666-d667c3d242a6" (UID: "034fe145-c83b-4b03-b666-d667c3d242a6"). InnerVolumeSpecName "kube-api-access-qx52p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.274115 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-scripts" (OuterVolumeSpecName: "scripts") pod "034fe145-c83b-4b03-b666-d667c3d242a6" (UID: "034fe145-c83b-4b03-b666-d667c3d242a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.349403 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "034fe145-c83b-4b03-b666-d667c3d242a6" (UID: "034fe145-c83b-4b03-b666-d667c3d242a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.371431 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.371460 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.371559 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/034fe145-c83b-4b03-b666-d667c3d242a6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.371568 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx52p\" (UniqueName: \"kubernetes.io/projected/034fe145-c83b-4b03-b666-d667c3d242a6-kube-api-access-qx52p\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.371580 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.378958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "034fe145-c83b-4b03-b666-d667c3d242a6" (UID: "034fe145-c83b-4b03-b666-d667c3d242a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.401007 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-config-data" (OuterVolumeSpecName: "config-data") pod "034fe145-c83b-4b03-b666-d667c3d242a6" (UID: "034fe145-c83b-4b03-b666-d667c3d242a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.473482 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:07 crc kubenswrapper[4809]: I1205 12:53:07.473517 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/034fe145-c83b-4b03-b666-d667c3d242a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.168317 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.213769 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.226846 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.243716 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:08 crc kubenswrapper[4809]: E1205 12:53:08.244308 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-central-agent" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244326 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-central-agent" Dec 05 12:53:08 crc kubenswrapper[4809]: E1205 12:53:08.244349 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="proxy-httpd" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244355 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="proxy-httpd" Dec 05 12:53:08 crc kubenswrapper[4809]: E1205 12:53:08.244373 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="sg-core" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244379 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="sg-core" Dec 05 12:53:08 crc kubenswrapper[4809]: E1205 12:53:08.244397 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-notification-agent" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244405 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-notification-agent" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244608 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-central-agent" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244653 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="sg-core" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244678 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="ceilometer-notification-agent" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.244688 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" containerName="proxy-httpd" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.247136 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.250449 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.250460 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.272993 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.320947 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-run-httpd\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.321108 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.321153 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-config-data\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.321269 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-log-httpd\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.321372 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-scripts\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.321420 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.321454 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcn72\" (UniqueName: \"kubernetes.io/projected/9f3bbc91-0344-4209-afce-c4bbf86d57e5-kube-api-access-qcn72\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.423757 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-run-httpd\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.423886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.423925 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-config-data\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.423962 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-log-httpd\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.424024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-scripts\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.424051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.424079 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcn72\" (UniqueName: \"kubernetes.io/projected/9f3bbc91-0344-4209-afce-c4bbf86d57e5-kube-api-access-qcn72\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.424767 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-log-httpd\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.425473 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-run-httpd\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.429374 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.431924 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-config-data\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.432187 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.434267 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-scripts\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.453351 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcn72\" (UniqueName: \"kubernetes.io/projected/9f3bbc91-0344-4209-afce-c4bbf86d57e5-kube-api-access-qcn72\") pod \"ceilometer-0\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.571995 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:08 crc kubenswrapper[4809]: I1205 12:53:08.898151 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="034fe145-c83b-4b03-b666-d667c3d242a6" path="/var/lib/kubelet/pods/034fe145-c83b-4b03-b666-d667c3d242a6/volumes" Dec 05 12:53:09 crc kubenswrapper[4809]: I1205 12:53:09.149597 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:09 crc kubenswrapper[4809]: W1205 12:53:09.164881 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f3bbc91_0344_4209_afce_c4bbf86d57e5.slice/crio-94939e936e780297d9bc60c7a6dd48cc2dd022082e6bc436e00e2accc8fa9f20 WatchSource:0}: Error finding container 94939e936e780297d9bc60c7a6dd48cc2dd022082e6bc436e00e2accc8fa9f20: Status 404 returned error can't find the container with id 94939e936e780297d9bc60c7a6dd48cc2dd022082e6bc436e00e2accc8fa9f20 Dec 05 12:53:09 crc kubenswrapper[4809]: I1205 12:53:09.187373 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerStarted","Data":"94939e936e780297d9bc60c7a6dd48cc2dd022082e6bc436e00e2accc8fa9f20"} Dec 05 12:53:10 crc kubenswrapper[4809]: I1205 12:53:10.198197 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerStarted","Data":"7f11b2f4549a9f5660b55cbfdc23c8d9ec3a62ecc94a118e7d61bab91b101822"} Dec 05 12:53:11 crc kubenswrapper[4809]: I1205 12:53:11.211579 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerStarted","Data":"db9f3934e6c47f63c835dc1d4a6514de73e84a5b863f752fe2bdd4e5eda3d40d"} Dec 05 12:53:12 crc kubenswrapper[4809]: I1205 12:53:12.222466 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerStarted","Data":"68c2affdb951a17d0ecc8e09b0a6b186fed08e407b2d61e0077ca57227b07be8"} Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.248244 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-tnb4v"] Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.250340 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.265571 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-tnb4v"] Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.354143 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79a7f59-b085-4378-8e8f-b28914905c8e-operator-scripts\") pod \"manila-db-create-tnb4v\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.354370 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8zt2\" (UniqueName: \"kubernetes.io/projected/f79a7f59-b085-4378-8e8f-b28914905c8e-kube-api-access-x8zt2\") pod \"manila-db-create-tnb4v\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.450188 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-e20e-account-create-update-f4ght"] Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.451734 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.453463 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.456137 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79a7f59-b085-4378-8e8f-b28914905c8e-operator-scripts\") pod \"manila-db-create-tnb4v\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.456270 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8zt2\" (UniqueName: \"kubernetes.io/projected/f79a7f59-b085-4378-8e8f-b28914905c8e-kube-api-access-x8zt2\") pod \"manila-db-create-tnb4v\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.457309 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79a7f59-b085-4378-8e8f-b28914905c8e-operator-scripts\") pod \"manila-db-create-tnb4v\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.469151 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-e20e-account-create-update-f4ght"] Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.489998 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8zt2\" (UniqueName: \"kubernetes.io/projected/f79a7f59-b085-4378-8e8f-b28914905c8e-kube-api-access-x8zt2\") pod \"manila-db-create-tnb4v\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.558299 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf396280-ba54-401d-b6ac-6aa8df85524f-operator-scripts\") pod \"manila-e20e-account-create-update-f4ght\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.558458 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6dtm\" (UniqueName: \"kubernetes.io/projected/bf396280-ba54-401d-b6ac-6aa8df85524f-kube-api-access-b6dtm\") pod \"manila-e20e-account-create-update-f4ght\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.568643 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.661391 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf396280-ba54-401d-b6ac-6aa8df85524f-operator-scripts\") pod \"manila-e20e-account-create-update-f4ght\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.662283 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf396280-ba54-401d-b6ac-6aa8df85524f-operator-scripts\") pod \"manila-e20e-account-create-update-f4ght\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.663031 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6dtm\" (UniqueName: \"kubernetes.io/projected/bf396280-ba54-401d-b6ac-6aa8df85524f-kube-api-access-b6dtm\") pod \"manila-e20e-account-create-update-f4ght\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.684331 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6dtm\" (UniqueName: \"kubernetes.io/projected/bf396280-ba54-401d-b6ac-6aa8df85524f-kube-api-access-b6dtm\") pod \"manila-e20e-account-create-update-f4ght\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:13 crc kubenswrapper[4809]: I1205 12:53:13.768164 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:14 crc kubenswrapper[4809]: W1205 12:53:14.181466 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf79a7f59_b085_4378_8e8f_b28914905c8e.slice/crio-5c91d6079ad2093b429e050b8fc3b0951ed8e7d3259faa8730685da979c3a6ea WatchSource:0}: Error finding container 5c91d6079ad2093b429e050b8fc3b0951ed8e7d3259faa8730685da979c3a6ea: Status 404 returned error can't find the container with id 5c91d6079ad2093b429e050b8fc3b0951ed8e7d3259faa8730685da979c3a6ea Dec 05 12:53:14 crc kubenswrapper[4809]: I1205 12:53:14.199534 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-tnb4v"] Dec 05 12:53:14 crc kubenswrapper[4809]: I1205 12:53:14.266416 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-tnb4v" event={"ID":"f79a7f59-b085-4378-8e8f-b28914905c8e","Type":"ContainerStarted","Data":"5c91d6079ad2093b429e050b8fc3b0951ed8e7d3259faa8730685da979c3a6ea"} Dec 05 12:53:14 crc kubenswrapper[4809]: I1205 12:53:14.279976 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerStarted","Data":"2ec8b12dcf0e0b5b614f35461eaf50a88984b406c1e69f8f38f59097988f3290"} Dec 05 12:53:14 crc kubenswrapper[4809]: I1205 12:53:14.281307 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 12:53:14 crc kubenswrapper[4809]: I1205 12:53:14.309333 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.446452624 podStartE2EDuration="6.309310265s" podCreationTimestamp="2025-12-05 12:53:08 +0000 UTC" firstStartedPulling="2025-12-05 12:53:09.172451806 +0000 UTC m=+6284.563428364" lastFinishedPulling="2025-12-05 12:53:13.035309447 +0000 UTC m=+6288.426286005" observedRunningTime="2025-12-05 12:53:14.299060019 +0000 UTC m=+6289.690036577" watchObservedRunningTime="2025-12-05 12:53:14.309310265 +0000 UTC m=+6289.700286833" Dec 05 12:53:14 crc kubenswrapper[4809]: I1205 12:53:14.333391 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-e20e-account-create-update-f4ght"] Dec 05 12:53:15 crc kubenswrapper[4809]: I1205 12:53:15.293261 4809 generic.go:334] "Generic (PLEG): container finished" podID="f79a7f59-b085-4378-8e8f-b28914905c8e" containerID="1e5d63d0c6bd6af6d27cab8a650b614c2e112c0ed70a541c0b81fa8e503eabaf" exitCode=0 Dec 05 12:53:15 crc kubenswrapper[4809]: I1205 12:53:15.293565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-tnb4v" event={"ID":"f79a7f59-b085-4378-8e8f-b28914905c8e","Type":"ContainerDied","Data":"1e5d63d0c6bd6af6d27cab8a650b614c2e112c0ed70a541c0b81fa8e503eabaf"} Dec 05 12:53:15 crc kubenswrapper[4809]: I1205 12:53:15.296732 4809 generic.go:334] "Generic (PLEG): container finished" podID="bf396280-ba54-401d-b6ac-6aa8df85524f" containerID="2cc76cd0c7d3848fa14b2566aa5d299351979a235ff9f92b1b232dd2131f325b" exitCode=0 Dec 05 12:53:15 crc kubenswrapper[4809]: I1205 12:53:15.297920 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e20e-account-create-update-f4ght" event={"ID":"bf396280-ba54-401d-b6ac-6aa8df85524f","Type":"ContainerDied","Data":"2cc76cd0c7d3848fa14b2566aa5d299351979a235ff9f92b1b232dd2131f325b"} Dec 05 12:53:15 crc kubenswrapper[4809]: I1205 12:53:15.297950 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e20e-account-create-update-f4ght" event={"ID":"bf396280-ba54-401d-b6ac-6aa8df85524f","Type":"ContainerStarted","Data":"dd15649cb1081748a55b24229f4e19e1a8fdc219849faf3ef641931a7df5418a"} Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.839282 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.846088 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.940133 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf396280-ba54-401d-b6ac-6aa8df85524f-operator-scripts\") pod \"bf396280-ba54-401d-b6ac-6aa8df85524f\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.940422 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8zt2\" (UniqueName: \"kubernetes.io/projected/f79a7f59-b085-4378-8e8f-b28914905c8e-kube-api-access-x8zt2\") pod \"f79a7f59-b085-4378-8e8f-b28914905c8e\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.940472 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6dtm\" (UniqueName: \"kubernetes.io/projected/bf396280-ba54-401d-b6ac-6aa8df85524f-kube-api-access-b6dtm\") pod \"bf396280-ba54-401d-b6ac-6aa8df85524f\" (UID: \"bf396280-ba54-401d-b6ac-6aa8df85524f\") " Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.940616 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79a7f59-b085-4378-8e8f-b28914905c8e-operator-scripts\") pod \"f79a7f59-b085-4378-8e8f-b28914905c8e\" (UID: \"f79a7f59-b085-4378-8e8f-b28914905c8e\") " Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.941170 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f79a7f59-b085-4378-8e8f-b28914905c8e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f79a7f59-b085-4378-8e8f-b28914905c8e" (UID: "f79a7f59-b085-4378-8e8f-b28914905c8e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.941346 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f79a7f59-b085-4378-8e8f-b28914905c8e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.942063 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf396280-ba54-401d-b6ac-6aa8df85524f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bf396280-ba54-401d-b6ac-6aa8df85524f" (UID: "bf396280-ba54-401d-b6ac-6aa8df85524f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.947599 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf396280-ba54-401d-b6ac-6aa8df85524f-kube-api-access-b6dtm" (OuterVolumeSpecName: "kube-api-access-b6dtm") pod "bf396280-ba54-401d-b6ac-6aa8df85524f" (UID: "bf396280-ba54-401d-b6ac-6aa8df85524f"). InnerVolumeSpecName "kube-api-access-b6dtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:53:16 crc kubenswrapper[4809]: I1205 12:53:16.947714 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f79a7f59-b085-4378-8e8f-b28914905c8e-kube-api-access-x8zt2" (OuterVolumeSpecName: "kube-api-access-x8zt2") pod "f79a7f59-b085-4378-8e8f-b28914905c8e" (UID: "f79a7f59-b085-4378-8e8f-b28914905c8e"). InnerVolumeSpecName "kube-api-access-x8zt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.043586 4809 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf396280-ba54-401d-b6ac-6aa8df85524f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.043647 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8zt2\" (UniqueName: \"kubernetes.io/projected/f79a7f59-b085-4378-8e8f-b28914905c8e-kube-api-access-x8zt2\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.043665 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6dtm\" (UniqueName: \"kubernetes.io/projected/bf396280-ba54-401d-b6ac-6aa8df85524f-kube-api-access-b6dtm\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.317732 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-tnb4v" event={"ID":"f79a7f59-b085-4378-8e8f-b28914905c8e","Type":"ContainerDied","Data":"5c91d6079ad2093b429e050b8fc3b0951ed8e7d3259faa8730685da979c3a6ea"} Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.317962 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c91d6079ad2093b429e050b8fc3b0951ed8e7d3259faa8730685da979c3a6ea" Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.318115 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-tnb4v" Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.320070 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e20e-account-create-update-f4ght" event={"ID":"bf396280-ba54-401d-b6ac-6aa8df85524f","Type":"ContainerDied","Data":"dd15649cb1081748a55b24229f4e19e1a8fdc219849faf3ef641931a7df5418a"} Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.320120 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd15649cb1081748a55b24229f4e19e1a8fdc219849faf3ef641931a7df5418a" Dec 05 12:53:17 crc kubenswrapper[4809]: I1205 12:53:17.320183 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e20e-account-create-update-f4ght" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.037943 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9kvm2"] Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.049066 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9kvm2"] Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.793012 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-r7h9j"] Dec 05 12:53:18 crc kubenswrapper[4809]: E1205 12:53:18.795269 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f79a7f59-b085-4378-8e8f-b28914905c8e" containerName="mariadb-database-create" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.795307 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f79a7f59-b085-4378-8e8f-b28914905c8e" containerName="mariadb-database-create" Dec 05 12:53:18 crc kubenswrapper[4809]: E1205 12:53:18.795360 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf396280-ba54-401d-b6ac-6aa8df85524f" containerName="mariadb-account-create-update" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.795373 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf396280-ba54-401d-b6ac-6aa8df85524f" containerName="mariadb-account-create-update" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.797178 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f79a7f59-b085-4378-8e8f-b28914905c8e" containerName="mariadb-database-create" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.797231 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf396280-ba54-401d-b6ac-6aa8df85524f" containerName="mariadb-account-create-update" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.799079 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.832077 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-6hww9" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.832303 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.842241 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-r7h9j"] Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.882809 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-job-config-data\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.883002 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-combined-ca-bundle\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.883075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-config-data\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.883115 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4tk\" (UniqueName: \"kubernetes.io/projected/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-kube-api-access-hk4tk\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.901148 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366869f6-f555-4b3b-95fe-e274e7d86652" path="/var/lib/kubelet/pods/366869f6-f555-4b3b-95fe-e274e7d86652/volumes" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.985172 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-job-config-data\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.985417 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-combined-ca-bundle\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.985496 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-config-data\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.985538 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4tk\" (UniqueName: \"kubernetes.io/projected/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-kube-api-access-hk4tk\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:18 crc kubenswrapper[4809]: I1205 12:53:18.991268 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-job-config-data\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:19 crc kubenswrapper[4809]: I1205 12:53:19.001553 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-config-data\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:19 crc kubenswrapper[4809]: I1205 12:53:19.004709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-combined-ca-bundle\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:19 crc kubenswrapper[4809]: I1205 12:53:19.007467 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4tk\" (UniqueName: \"kubernetes.io/projected/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-kube-api-access-hk4tk\") pod \"manila-db-sync-r7h9j\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:19 crc kubenswrapper[4809]: I1205 12:53:19.159223 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:19 crc kubenswrapper[4809]: I1205 12:53:19.768913 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-r7h9j"] Dec 05 12:53:20 crc kubenswrapper[4809]: I1205 12:53:20.368645 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-r7h9j" event={"ID":"4db0f5c6-097d-4923-be05-a74f8ddb5fbe","Type":"ContainerStarted","Data":"34e70d3d2a2b21c056e2fd176eaa068b3d41322fe1c4cecc8a01221c1815a088"} Dec 05 12:53:25 crc kubenswrapper[4809]: I1205 12:53:25.423820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-r7h9j" event={"ID":"4db0f5c6-097d-4923-be05-a74f8ddb5fbe","Type":"ContainerStarted","Data":"4be7f636c4c494a709b8c3fa18951559d0d867fbc0e6453d63cfb10c93df2c52"} Dec 05 12:53:25 crc kubenswrapper[4809]: I1205 12:53:25.455936 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-r7h9j" podStartSLOduration=3.346549547 podStartE2EDuration="7.45591615s" podCreationTimestamp="2025-12-05 12:53:18 +0000 UTC" firstStartedPulling="2025-12-05 12:53:19.776973146 +0000 UTC m=+6295.167949704" lastFinishedPulling="2025-12-05 12:53:23.886339749 +0000 UTC m=+6299.277316307" observedRunningTime="2025-12-05 12:53:25.440230828 +0000 UTC m=+6300.831207426" watchObservedRunningTime="2025-12-05 12:53:25.45591615 +0000 UTC m=+6300.846892728" Dec 05 12:53:27 crc kubenswrapper[4809]: I1205 12:53:27.453315 4809 generic.go:334] "Generic (PLEG): container finished" podID="4db0f5c6-097d-4923-be05-a74f8ddb5fbe" containerID="4be7f636c4c494a709b8c3fa18951559d0d867fbc0e6453d63cfb10c93df2c52" exitCode=0 Dec 05 12:53:27 crc kubenswrapper[4809]: I1205 12:53:27.453442 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-r7h9j" event={"ID":"4db0f5c6-097d-4923-be05-a74f8ddb5fbe","Type":"ContainerDied","Data":"4be7f636c4c494a709b8c3fa18951559d0d867fbc0e6453d63cfb10c93df2c52"} Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.052602 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.219297 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk4tk\" (UniqueName: \"kubernetes.io/projected/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-kube-api-access-hk4tk\") pod \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.219494 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-config-data\") pod \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.219601 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-combined-ca-bundle\") pod \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.219861 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-job-config-data\") pod \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\" (UID: \"4db0f5c6-097d-4923-be05-a74f8ddb5fbe\") " Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.226806 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "4db0f5c6-097d-4923-be05-a74f8ddb5fbe" (UID: "4db0f5c6-097d-4923-be05-a74f8ddb5fbe"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.227081 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-kube-api-access-hk4tk" (OuterVolumeSpecName: "kube-api-access-hk4tk") pod "4db0f5c6-097d-4923-be05-a74f8ddb5fbe" (UID: "4db0f5c6-097d-4923-be05-a74f8ddb5fbe"). InnerVolumeSpecName "kube-api-access-hk4tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.233769 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-config-data" (OuterVolumeSpecName: "config-data") pod "4db0f5c6-097d-4923-be05-a74f8ddb5fbe" (UID: "4db0f5c6-097d-4923-be05-a74f8ddb5fbe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.273147 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4db0f5c6-097d-4923-be05-a74f8ddb5fbe" (UID: "4db0f5c6-097d-4923-be05-a74f8ddb5fbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.324355 4809 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.324432 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk4tk\" (UniqueName: \"kubernetes.io/projected/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-kube-api-access-hk4tk\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.324468 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.324494 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db0f5c6-097d-4923-be05-a74f8ddb5fbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.480098 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-r7h9j" event={"ID":"4db0f5c6-097d-4923-be05-a74f8ddb5fbe","Type":"ContainerDied","Data":"34e70d3d2a2b21c056e2fd176eaa068b3d41322fe1c4cecc8a01221c1815a088"} Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.480163 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34e70d3d2a2b21c056e2fd176eaa068b3d41322fe1c4cecc8a01221c1815a088" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.480115 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-r7h9j" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.879359 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 12:53:29 crc kubenswrapper[4809]: E1205 12:53:29.880141 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db0f5c6-097d-4923-be05-a74f8ddb5fbe" containerName="manila-db-sync" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.880156 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db0f5c6-097d-4923-be05-a74f8ddb5fbe" containerName="manila-db-sync" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.880402 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db0f5c6-097d-4923-be05-a74f8ddb5fbe" containerName="manila-db-sync" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.881607 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.890817 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.890981 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-6hww9" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.891108 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.892382 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.895965 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.913891 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.913677 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.922652 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.933033 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.939930 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6747d5f4b9-lmbf9"] Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.941907 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:29 crc kubenswrapper[4809]: I1205 12:53:29.964476 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6747d5f4b9-lmbf9"] Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.043412 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eec2e149-075e-4b45-9140-2eba7beef8c4-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.043738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/575f4d22-5809-4ca1-bbf7-fdbdbe305194-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.043866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-config-data\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.043942 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.044051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgkhk\" (UniqueName: \"kubernetes.io/projected/eec2e149-075e-4b45-9140-2eba7beef8c4-kube-api-access-zgkhk\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.044167 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljlc5\" (UniqueName: \"kubernetes.io/projected/be4ba468-19eb-4918-8fb6-cf9653e6db32-kube-api-access-ljlc5\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.044273 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-dns-svc\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.044373 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm99f\" (UniqueName: \"kubernetes.io/projected/575f4d22-5809-4ca1-bbf7-fdbdbe305194-kube-api-access-rm99f\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.044508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-scripts\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.044685 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.044872 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-config\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.045001 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/575f4d22-5809-4ca1-bbf7-fdbdbe305194-ceph\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.045112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.045237 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/575f4d22-5809-4ca1-bbf7-fdbdbe305194-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.045413 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-scripts\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.045540 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-config-data\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.045700 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-sb\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.046156 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.046309 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-nb\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.078274 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.119053 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.119211 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.122352 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150340 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-nb\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150399 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eec2e149-075e-4b45-9140-2eba7beef8c4-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150427 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/575f4d22-5809-4ca1-bbf7-fdbdbe305194-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150451 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-config-data\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150470 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgkhk\" (UniqueName: \"kubernetes.io/projected/eec2e149-075e-4b45-9140-2eba7beef8c4-kube-api-access-zgkhk\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150521 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljlc5\" (UniqueName: \"kubernetes.io/projected/be4ba468-19eb-4918-8fb6-cf9653e6db32-kube-api-access-ljlc5\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150539 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-dns-svc\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm99f\" (UniqueName: \"kubernetes.io/projected/575f4d22-5809-4ca1-bbf7-fdbdbe305194-kube-api-access-rm99f\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-scripts\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150655 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-config\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/575f4d22-5809-4ca1-bbf7-fdbdbe305194-ceph\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150690 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150710 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/575f4d22-5809-4ca1-bbf7-fdbdbe305194-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150750 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-scripts\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150770 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-config-data\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150793 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-sb\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.150817 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.152927 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-nb\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.153004 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eec2e149-075e-4b45-9140-2eba7beef8c4-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.153042 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/575f4d22-5809-4ca1-bbf7-fdbdbe305194-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.155197 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-dns-svc\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.156696 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.156800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/575f4d22-5809-4ca1-bbf7-fdbdbe305194-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.158261 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-config-data\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.158348 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.158484 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-config\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.160282 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.162016 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-scripts\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.164385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-sb\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.169358 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/575f4d22-5809-4ca1-bbf7-fdbdbe305194-ceph\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.170922 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.171834 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm99f\" (UniqueName: \"kubernetes.io/projected/575f4d22-5809-4ca1-bbf7-fdbdbe305194-kube-api-access-rm99f\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.172568 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljlc5\" (UniqueName: \"kubernetes.io/projected/be4ba468-19eb-4918-8fb6-cf9653e6db32-kube-api-access-ljlc5\") pod \"dnsmasq-dns-6747d5f4b9-lmbf9\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.173168 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec2e149-075e-4b45-9140-2eba7beef8c4-scripts\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.173962 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgkhk\" (UniqueName: \"kubernetes.io/projected/eec2e149-075e-4b45-9140-2eba7beef8c4-kube-api-access-zgkhk\") pod \"manila-scheduler-0\" (UID: \"eec2e149-075e-4b45-9140-2eba7beef8c4\") " pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.183335 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575f4d22-5809-4ca1-bbf7-fdbdbe305194-config-data\") pod \"manila-share-share1-0\" (UID: \"575f4d22-5809-4ca1-bbf7-fdbdbe305194\") " pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.217232 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.246428 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.252084 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb7f996f-ece9-4cfc-9744-f331e75550e4-etc-machine-id\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.252146 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-config-data\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.252165 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-config-data-custom\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.252193 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-scripts\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.252238 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.252254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5dmd\" (UniqueName: \"kubernetes.io/projected/fb7f996f-ece9-4cfc-9744-f331e75550e4-kube-api-access-z5dmd\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.252270 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7f996f-ece9-4cfc-9744-f331e75550e4-logs\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.265414 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.354329 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.354736 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5dmd\" (UniqueName: \"kubernetes.io/projected/fb7f996f-ece9-4cfc-9744-f331e75550e4-kube-api-access-z5dmd\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.354764 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7f996f-ece9-4cfc-9744-f331e75550e4-logs\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.354935 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb7f996f-ece9-4cfc-9744-f331e75550e4-etc-machine-id\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.354996 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-config-data\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.355022 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-config-data-custom\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.355058 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-scripts\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.355394 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb7f996f-ece9-4cfc-9744-f331e75550e4-logs\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.355463 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb7f996f-ece9-4cfc-9744-f331e75550e4-etc-machine-id\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.361855 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-scripts\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.363899 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-config-data\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.372308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.381647 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5dmd\" (UniqueName: \"kubernetes.io/projected/fb7f996f-ece9-4cfc-9744-f331e75550e4-kube-api-access-z5dmd\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.390569 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb7f996f-ece9-4cfc-9744-f331e75550e4-config-data-custom\") pod \"manila-api-0\" (UID: \"fb7f996f-ece9-4cfc-9744-f331e75550e4\") " pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.458282 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.803296 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 12:53:30 crc kubenswrapper[4809]: I1205 12:53:30.993550 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6747d5f4b9-lmbf9"] Dec 05 12:53:31 crc kubenswrapper[4809]: I1205 12:53:31.216335 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 12:53:31 crc kubenswrapper[4809]: I1205 12:53:31.369730 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 12:53:31 crc kubenswrapper[4809]: I1205 12:53:31.536995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eec2e149-075e-4b45-9140-2eba7beef8c4","Type":"ContainerStarted","Data":"4b8d3ab3965df08e4e233d9132aab7280168ad5ec3ffe005c10eea317ae5bad3"} Dec 05 12:53:31 crc kubenswrapper[4809]: I1205 12:53:31.537998 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" event={"ID":"be4ba468-19eb-4918-8fb6-cf9653e6db32","Type":"ContainerStarted","Data":"ee1969a7035f914ccb31f71c6e72620c7e3a245410796956e9b670cf4977ebab"} Dec 05 12:53:31 crc kubenswrapper[4809]: I1205 12:53:31.538801 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"575f4d22-5809-4ca1-bbf7-fdbdbe305194","Type":"ContainerStarted","Data":"92814c3731ce60b01bacdf179d421141c5cbe85c1aed3aa97fcc375de42fa89c"} Dec 05 12:53:31 crc kubenswrapper[4809]: W1205 12:53:31.989592 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb7f996f_ece9_4cfc_9744_f331e75550e4.slice/crio-90fa6792e0d587cc72735eea343e6b3a1062dc97fe52c5f156c9c19a68052919 WatchSource:0}: Error finding container 90fa6792e0d587cc72735eea343e6b3a1062dc97fe52c5f156c9c19a68052919: Status 404 returned error can't find the container with id 90fa6792e0d587cc72735eea343e6b3a1062dc97fe52c5f156c9c19a68052919 Dec 05 12:53:32 crc kubenswrapper[4809]: I1205 12:53:32.568807 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" event={"ID":"be4ba468-19eb-4918-8fb6-cf9653e6db32","Type":"ContainerDied","Data":"0ce47b66788d185a474a5dc32dc73ca616112c316062f13a8603608063800aab"} Dec 05 12:53:32 crc kubenswrapper[4809]: I1205 12:53:32.568763 4809 generic.go:334] "Generic (PLEG): container finished" podID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerID="0ce47b66788d185a474a5dc32dc73ca616112c316062f13a8603608063800aab" exitCode=0 Dec 05 12:53:32 crc kubenswrapper[4809]: I1205 12:53:32.593152 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"fb7f996f-ece9-4cfc-9744-f331e75550e4","Type":"ContainerStarted","Data":"90fa6792e0d587cc72735eea343e6b3a1062dc97fe52c5f156c9c19a68052919"} Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.605538 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" event={"ID":"be4ba468-19eb-4918-8fb6-cf9653e6db32","Type":"ContainerStarted","Data":"868c37abc97d13f339f7a35d70e8a52d14039ff85a0f002a0ba80ba270e35204"} Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.606017 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.610984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"fb7f996f-ece9-4cfc-9744-f331e75550e4","Type":"ContainerStarted","Data":"7d0906a3429ed8f72d276c07c78b32110eec8a3a6e826c7687081ed8e6a36243"} Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.611031 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"fb7f996f-ece9-4cfc-9744-f331e75550e4","Type":"ContainerStarted","Data":"e0d7cd301e2dbefec5074c9179c6a30f431d3d87ae24cd542557f54f2b5e5bdc"} Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.611107 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.613862 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eec2e149-075e-4b45-9140-2eba7beef8c4","Type":"ContainerStarted","Data":"2dfbaeec4bc65ffbeeb4e6866e01398f45875cb05306fa61394af5eca95f8e6a"} Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.613927 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eec2e149-075e-4b45-9140-2eba7beef8c4","Type":"ContainerStarted","Data":"cb6b05d0b5f3034e966e1a503eacc2e2e7037e34e76ac856de2e76c80dd29bb9"} Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.635624 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" podStartSLOduration=4.635601649 podStartE2EDuration="4.635601649s" podCreationTimestamp="2025-12-05 12:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:53:33.626237197 +0000 UTC m=+6309.017213755" watchObservedRunningTime="2025-12-05 12:53:33.635601649 +0000 UTC m=+6309.026578207" Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.657459 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.657435366 podStartE2EDuration="3.657435366s" podCreationTimestamp="2025-12-05 12:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:53:33.643279566 +0000 UTC m=+6309.034256114" watchObservedRunningTime="2025-12-05 12:53:33.657435366 +0000 UTC m=+6309.048411924" Dec 05 12:53:33 crc kubenswrapper[4809]: I1205 12:53:33.664572 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.450872393 podStartE2EDuration="4.664555678s" podCreationTimestamp="2025-12-05 12:53:29 +0000 UTC" firstStartedPulling="2025-12-05 12:53:30.804388933 +0000 UTC m=+6306.195365491" lastFinishedPulling="2025-12-05 12:53:32.018072218 +0000 UTC m=+6307.409048776" observedRunningTime="2025-12-05 12:53:33.662561114 +0000 UTC m=+6309.053537682" watchObservedRunningTime="2025-12-05 12:53:33.664555678 +0000 UTC m=+6309.055532226" Dec 05 12:53:38 crc kubenswrapper[4809]: I1205 12:53:38.589902 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 12:53:39 crc kubenswrapper[4809]: I1205 12:53:39.707117 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"575f4d22-5809-4ca1-bbf7-fdbdbe305194","Type":"ContainerStarted","Data":"309ade206b61ab43564b9ff6360282ee1592c935680db9855e1e8ed6efa1e593"} Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.218035 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.266849 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.356528 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6974cc94df-frb66"] Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.356770 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6974cc94df-frb66" podUID="9da25051-70b9-4014-b10c-900022a35a1a" containerName="dnsmasq-dns" containerID="cri-o://ddcbe5ce7b0f365a6b07d0da94b0da1a31dbf33569aebaa430d5cae3be051a4c" gracePeriod=10 Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.729701 4809 generic.go:334] "Generic (PLEG): container finished" podID="9da25051-70b9-4014-b10c-900022a35a1a" containerID="ddcbe5ce7b0f365a6b07d0da94b0da1a31dbf33569aebaa430d5cae3be051a4c" exitCode=0 Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.729984 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6974cc94df-frb66" event={"ID":"9da25051-70b9-4014-b10c-900022a35a1a","Type":"ContainerDied","Data":"ddcbe5ce7b0f365a6b07d0da94b0da1a31dbf33569aebaa430d5cae3be051a4c"} Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.731232 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"575f4d22-5809-4ca1-bbf7-fdbdbe305194","Type":"ContainerStarted","Data":"3432499fdf852bc2198ed2cedd76306825f2fb4f80939cfacf7320a6930260cd"} Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.769998 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.067905244 podStartE2EDuration="11.769977522s" podCreationTimestamp="2025-12-05 12:53:29 +0000 UTC" firstStartedPulling="2025-12-05 12:53:31.242760688 +0000 UTC m=+6306.633737246" lastFinishedPulling="2025-12-05 12:53:38.944832966 +0000 UTC m=+6314.335809524" observedRunningTime="2025-12-05 12:53:40.768549094 +0000 UTC m=+6316.159525672" watchObservedRunningTime="2025-12-05 12:53:40.769977522 +0000 UTC m=+6316.160954090" Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.933137 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.996670 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrkqv\" (UniqueName: \"kubernetes.io/projected/9da25051-70b9-4014-b10c-900022a35a1a-kube-api-access-vrkqv\") pod \"9da25051-70b9-4014-b10c-900022a35a1a\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.996781 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-dns-svc\") pod \"9da25051-70b9-4014-b10c-900022a35a1a\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.996847 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-config\") pod \"9da25051-70b9-4014-b10c-900022a35a1a\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.996987 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-sb\") pod \"9da25051-70b9-4014-b10c-900022a35a1a\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " Dec 05 12:53:40 crc kubenswrapper[4809]: I1205 12:53:40.997079 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-nb\") pod \"9da25051-70b9-4014-b10c-900022a35a1a\" (UID: \"9da25051-70b9-4014-b10c-900022a35a1a\") " Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.007411 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9da25051-70b9-4014-b10c-900022a35a1a-kube-api-access-vrkqv" (OuterVolumeSpecName: "kube-api-access-vrkqv") pod "9da25051-70b9-4014-b10c-900022a35a1a" (UID: "9da25051-70b9-4014-b10c-900022a35a1a"). InnerVolumeSpecName "kube-api-access-vrkqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.092704 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9da25051-70b9-4014-b10c-900022a35a1a" (UID: "9da25051-70b9-4014-b10c-900022a35a1a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.095647 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9da25051-70b9-4014-b10c-900022a35a1a" (UID: "9da25051-70b9-4014-b10c-900022a35a1a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.100597 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.100662 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.100675 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrkqv\" (UniqueName: \"kubernetes.io/projected/9da25051-70b9-4014-b10c-900022a35a1a-kube-api-access-vrkqv\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.107724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9da25051-70b9-4014-b10c-900022a35a1a" (UID: "9da25051-70b9-4014-b10c-900022a35a1a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.112569 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-config" (OuterVolumeSpecName: "config") pod "9da25051-70b9-4014-b10c-900022a35a1a" (UID: "9da25051-70b9-4014-b10c-900022a35a1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.203937 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.204078 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da25051-70b9-4014-b10c-900022a35a1a-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.743929 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6974cc94df-frb66" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.746095 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6974cc94df-frb66" event={"ID":"9da25051-70b9-4014-b10c-900022a35a1a","Type":"ContainerDied","Data":"1ed84707883bb40c4f407034fbec26fdd29d63025ba2b3164379aaf6a06f5729"} Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.746144 4809 scope.go:117] "RemoveContainer" containerID="ddcbe5ce7b0f365a6b07d0da94b0da1a31dbf33569aebaa430d5cae3be051a4c" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.822750 4809 scope.go:117] "RemoveContainer" containerID="0c311f82c9edb7b608b237a67d4bd4e751e5f1ce3fd4a92d8f4043b66d73021e" Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.828802 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6974cc94df-frb66"] Dec 05 12:53:41 crc kubenswrapper[4809]: I1205 12:53:41.850400 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6974cc94df-frb66"] Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.541567 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.542224 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-central-agent" containerID="cri-o://7f11b2f4549a9f5660b55cbfdc23c8d9ec3a62ecc94a118e7d61bab91b101822" gracePeriod=30 Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.542666 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-notification-agent" containerID="cri-o://db9f3934e6c47f63c835dc1d4a6514de73e84a5b863f752fe2bdd4e5eda3d40d" gracePeriod=30 Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.542689 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="sg-core" containerID="cri-o://68c2affdb951a17d0ecc8e09b0a6b186fed08e407b2d61e0077ca57227b07be8" gracePeriod=30 Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.542846 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="proxy-httpd" containerID="cri-o://2ec8b12dcf0e0b5b614f35461eaf50a88984b406c1e69f8f38f59097988f3290" gracePeriod=30 Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.757427 4809 generic.go:334] "Generic (PLEG): container finished" podID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerID="68c2affdb951a17d0ecc8e09b0a6b186fed08e407b2d61e0077ca57227b07be8" exitCode=2 Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.757489 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerDied","Data":"68c2affdb951a17d0ecc8e09b0a6b186fed08e407b2d61e0077ca57227b07be8"} Dec 05 12:53:42 crc kubenswrapper[4809]: I1205 12:53:42.885350 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9da25051-70b9-4014-b10c-900022a35a1a" path="/var/lib/kubelet/pods/9da25051-70b9-4014-b10c-900022a35a1a/volumes" Dec 05 12:53:43 crc kubenswrapper[4809]: I1205 12:53:43.770352 4809 generic.go:334] "Generic (PLEG): container finished" podID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerID="2ec8b12dcf0e0b5b614f35461eaf50a88984b406c1e69f8f38f59097988f3290" exitCode=0 Dec 05 12:53:43 crc kubenswrapper[4809]: I1205 12:53:43.770598 4809 generic.go:334] "Generic (PLEG): container finished" podID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerID="7f11b2f4549a9f5660b55cbfdc23c8d9ec3a62ecc94a118e7d61bab91b101822" exitCode=0 Dec 05 12:53:43 crc kubenswrapper[4809]: I1205 12:53:43.770616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerDied","Data":"2ec8b12dcf0e0b5b614f35461eaf50a88984b406c1e69f8f38f59097988f3290"} Dec 05 12:53:43 crc kubenswrapper[4809]: I1205 12:53:43.770658 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerDied","Data":"7f11b2f4549a9f5660b55cbfdc23c8d9ec3a62ecc94a118e7d61bab91b101822"} Dec 05 12:53:43 crc kubenswrapper[4809]: I1205 12:53:43.945975 4809 scope.go:117] "RemoveContainer" containerID="0c7adc9e5c2b4683aee4eee140d398f170c888a8277719398a288127039103c6" Dec 05 12:53:43 crc kubenswrapper[4809]: I1205 12:53:43.981302 4809 scope.go:117] "RemoveContainer" containerID="88013cfb304725872ba8f39d1287ad224065d6e1ffe1ded2a37d03ae03693e1d" Dec 05 12:53:44 crc kubenswrapper[4809]: I1205 12:53:44.037099 4809 scope.go:117] "RemoveContainer" containerID="c39dc3f714c42f6c61812c5362fbb6e3bfded98b503aa52a548067b44d180c1e" Dec 05 12:53:44 crc kubenswrapper[4809]: I1205 12:53:44.047096 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:53:44 crc kubenswrapper[4809]: I1205 12:53:44.047187 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:53:44 crc kubenswrapper[4809]: I1205 12:53:44.067240 4809 scope.go:117] "RemoveContainer" containerID="5f97f36da62aa79461a6d927b851869fe25656baf29470378972b7cc29f31157" Dec 05 12:53:47 crc kubenswrapper[4809]: I1205 12:53:47.814207 4809 generic.go:334] "Generic (PLEG): container finished" podID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerID="db9f3934e6c47f63c835dc1d4a6514de73e84a5b863f752fe2bdd4e5eda3d40d" exitCode=0 Dec 05 12:53:47 crc kubenswrapper[4809]: I1205 12:53:47.814289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerDied","Data":"db9f3934e6c47f63c835dc1d4a6514de73e84a5b863f752fe2bdd4e5eda3d40d"} Dec 05 12:53:47 crc kubenswrapper[4809]: I1205 12:53:47.963007 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.055435 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-scripts\") pod \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.055504 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-combined-ca-bundle\") pod \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.056142 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-config-data\") pod \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.056270 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-run-httpd\") pod \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.056432 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-sg-core-conf-yaml\") pod \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.056779 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9f3bbc91-0344-4209-afce-c4bbf86d57e5" (UID: "9f3bbc91-0344-4209-afce-c4bbf86d57e5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.056856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcn72\" (UniqueName: \"kubernetes.io/projected/9f3bbc91-0344-4209-afce-c4bbf86d57e5-kube-api-access-qcn72\") pod \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.058170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-log-httpd\") pod \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\" (UID: \"9f3bbc91-0344-4209-afce-c4bbf86d57e5\") " Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.058924 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9f3bbc91-0344-4209-afce-c4bbf86d57e5" (UID: "9f3bbc91-0344-4209-afce-c4bbf86d57e5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.060243 4809 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.060299 4809 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f3bbc91-0344-4209-afce-c4bbf86d57e5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.063383 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-scripts" (OuterVolumeSpecName: "scripts") pod "9f3bbc91-0344-4209-afce-c4bbf86d57e5" (UID: "9f3bbc91-0344-4209-afce-c4bbf86d57e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.078855 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f3bbc91-0344-4209-afce-c4bbf86d57e5-kube-api-access-qcn72" (OuterVolumeSpecName: "kube-api-access-qcn72") pod "9f3bbc91-0344-4209-afce-c4bbf86d57e5" (UID: "9f3bbc91-0344-4209-afce-c4bbf86d57e5"). InnerVolumeSpecName "kube-api-access-qcn72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.107893 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9f3bbc91-0344-4209-afce-c4bbf86d57e5" (UID: "9f3bbc91-0344-4209-afce-c4bbf86d57e5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.162083 4809 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.162113 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcn72\" (UniqueName: \"kubernetes.io/projected/9f3bbc91-0344-4209-afce-c4bbf86d57e5-kube-api-access-qcn72\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.162127 4809 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.182954 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f3bbc91-0344-4209-afce-c4bbf86d57e5" (UID: "9f3bbc91-0344-4209-afce-c4bbf86d57e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.191828 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-config-data" (OuterVolumeSpecName: "config-data") pod "9f3bbc91-0344-4209-afce-c4bbf86d57e5" (UID: "9f3bbc91-0344-4209-afce-c4bbf86d57e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.263943 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.264208 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3bbc91-0344-4209-afce-c4bbf86d57e5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.829024 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f3bbc91-0344-4209-afce-c4bbf86d57e5","Type":"ContainerDied","Data":"94939e936e780297d9bc60c7a6dd48cc2dd022082e6bc436e00e2accc8fa9f20"} Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.829151 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.829386 4809 scope.go:117] "RemoveContainer" containerID="2ec8b12dcf0e0b5b614f35461eaf50a88984b406c1e69f8f38f59097988f3290" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.865850 4809 scope.go:117] "RemoveContainer" containerID="68c2affdb951a17d0ecc8e09b0a6b186fed08e407b2d61e0077ca57227b07be8" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.913356 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.919647 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.931745 4809 scope.go:117] "RemoveContainer" containerID="db9f3934e6c47f63c835dc1d4a6514de73e84a5b863f752fe2bdd4e5eda3d40d" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.941809 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:48 crc kubenswrapper[4809]: E1205 12:53:48.942710 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9da25051-70b9-4014-b10c-900022a35a1a" containerName="dnsmasq-dns" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.942838 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9da25051-70b9-4014-b10c-900022a35a1a" containerName="dnsmasq-dns" Dec 05 12:53:48 crc kubenswrapper[4809]: E1205 12:53:48.943019 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="proxy-httpd" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.943119 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="proxy-httpd" Dec 05 12:53:48 crc kubenswrapper[4809]: E1205 12:53:48.943250 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="sg-core" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.943342 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="sg-core" Dec 05 12:53:48 crc kubenswrapper[4809]: E1205 12:53:48.943441 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9da25051-70b9-4014-b10c-900022a35a1a" containerName="init" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.943541 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9da25051-70b9-4014-b10c-900022a35a1a" containerName="init" Dec 05 12:53:48 crc kubenswrapper[4809]: E1205 12:53:48.943694 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-central-agent" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.943797 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-central-agent" Dec 05 12:53:48 crc kubenswrapper[4809]: E1205 12:53:48.943917 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-notification-agent" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.944025 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-notification-agent" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.944512 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-central-agent" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.944623 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9da25051-70b9-4014-b10c-900022a35a1a" containerName="dnsmasq-dns" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.944797 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="ceilometer-notification-agent" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.944919 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="proxy-httpd" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.945072 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" containerName="sg-core" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.948360 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.951034 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.951331 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.951955 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.958846 4809 scope.go:117] "RemoveContainer" containerID="7f11b2f4549a9f5660b55cbfdc23c8d9ec3a62ecc94a118e7d61bab91b101822" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.986462 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63f6c3ad-0465-4149-98ae-da47e309692d-log-httpd\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.986564 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-config-data\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.986614 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63f6c3ad-0465-4149-98ae-da47e309692d-run-httpd\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.986715 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.986759 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.987396 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-scripts\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:48 crc kubenswrapper[4809]: I1205 12:53:48.987494 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8htp\" (UniqueName: \"kubernetes.io/projected/63f6c3ad-0465-4149-98ae-da47e309692d-kube-api-access-h8htp\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.090213 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8htp\" (UniqueName: \"kubernetes.io/projected/63f6c3ad-0465-4149-98ae-da47e309692d-kube-api-access-h8htp\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.090363 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63f6c3ad-0465-4149-98ae-da47e309692d-log-httpd\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.090396 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-config-data\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.090420 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63f6c3ad-0465-4149-98ae-da47e309692d-run-httpd\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.090452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.090479 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.090515 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-scripts\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.091859 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63f6c3ad-0465-4149-98ae-da47e309692d-run-httpd\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.092115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63f6c3ad-0465-4149-98ae-da47e309692d-log-httpd\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.095357 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-scripts\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.095657 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.099345 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.099983 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63f6c3ad-0465-4149-98ae-da47e309692d-config-data\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.110114 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8htp\" (UniqueName: \"kubernetes.io/projected/63f6c3ad-0465-4149-98ae-da47e309692d-kube-api-access-h8htp\") pod \"ceilometer-0\" (UID: \"63f6c3ad-0465-4149-98ae-da47e309692d\") " pod="openstack/ceilometer-0" Dec 05 12:53:49 crc kubenswrapper[4809]: I1205 12:53:49.287392 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 12:53:50 crc kubenswrapper[4809]: W1205 12:53:50.023858 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63f6c3ad_0465_4149_98ae_da47e309692d.slice/crio-8bf862fd28ec8a5b1a408ef92e519a86d138589de41d6504501c608327179073 WatchSource:0}: Error finding container 8bf862fd28ec8a5b1a408ef92e519a86d138589de41d6504501c608327179073: Status 404 returned error can't find the container with id 8bf862fd28ec8a5b1a408ef92e519a86d138589de41d6504501c608327179073 Dec 05 12:53:50 crc kubenswrapper[4809]: I1205 12:53:50.031367 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 12:53:50 crc kubenswrapper[4809]: I1205 12:53:50.247761 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 05 12:53:50 crc kubenswrapper[4809]: I1205 12:53:50.857186 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63f6c3ad-0465-4149-98ae-da47e309692d","Type":"ContainerStarted","Data":"9e711cfad60fbbac52f59c07487d86e3b2850014368ff2e5b0ddeb8bfbf39511"} Dec 05 12:53:50 crc kubenswrapper[4809]: I1205 12:53:50.857868 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63f6c3ad-0465-4149-98ae-da47e309692d","Type":"ContainerStarted","Data":"8bf862fd28ec8a5b1a408ef92e519a86d138589de41d6504501c608327179073"} Dec 05 12:53:50 crc kubenswrapper[4809]: I1205 12:53:50.885642 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f3bbc91-0344-4209-afce-c4bbf86d57e5" path="/var/lib/kubelet/pods/9f3bbc91-0344-4209-afce-c4bbf86d57e5/volumes" Dec 05 12:53:51 crc kubenswrapper[4809]: I1205 12:53:51.855703 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 05 12:53:51 crc kubenswrapper[4809]: I1205 12:53:51.871133 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63f6c3ad-0465-4149-98ae-da47e309692d","Type":"ContainerStarted","Data":"92c00e8014e492ad7bb391cca7fc21df95e5ae6bd8fdd87147ba6540ef6ad27e"} Dec 05 12:53:51 crc kubenswrapper[4809]: I1205 12:53:51.881076 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 05 12:53:51 crc kubenswrapper[4809]: I1205 12:53:51.965651 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 05 12:53:52 crc kubenswrapper[4809]: I1205 12:53:52.889844 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63f6c3ad-0465-4149-98ae-da47e309692d","Type":"ContainerStarted","Data":"8dc6af80d350598e04fd905803380f8acc7557520f93fa96e4582fc9a1c28018"} Dec 05 12:53:53 crc kubenswrapper[4809]: I1205 12:53:53.899518 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63f6c3ad-0465-4149-98ae-da47e309692d","Type":"ContainerStarted","Data":"04daeb15c69aaa10e6e4232bf0dbd6a9e162b60fe2deb534cbde046479f360f1"} Dec 05 12:53:53 crc kubenswrapper[4809]: I1205 12:53:53.900074 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 12:53:59 crc kubenswrapper[4809]: I1205 12:53:59.037663 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=7.98311987 podStartE2EDuration="11.037617432s" podCreationTimestamp="2025-12-05 12:53:48 +0000 UTC" firstStartedPulling="2025-12-05 12:53:50.026455172 +0000 UTC m=+6325.417431740" lastFinishedPulling="2025-12-05 12:53:53.080952734 +0000 UTC m=+6328.471929302" observedRunningTime="2025-12-05 12:53:53.923701459 +0000 UTC m=+6329.314678027" watchObservedRunningTime="2025-12-05 12:53:59.037617432 +0000 UTC m=+6334.428594010" Dec 05 12:53:59 crc kubenswrapper[4809]: I1205 12:53:59.047230 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7e82-account-create-update-xwx56"] Dec 05 12:53:59 crc kubenswrapper[4809]: I1205 12:53:59.060505 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7e82-account-create-update-xwx56"] Dec 05 12:53:59 crc kubenswrapper[4809]: I1205 12:53:59.073022 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wj5vp"] Dec 05 12:53:59 crc kubenswrapper[4809]: I1205 12:53:59.086100 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wj5vp"] Dec 05 12:54:00 crc kubenswrapper[4809]: I1205 12:54:00.884838 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0f9da55-e8ce-4c11-acd6-b3193cdb660d" path="/var/lib/kubelet/pods/c0f9da55-e8ce-4c11-acd6-b3193cdb660d/volumes" Dec 05 12:54:00 crc kubenswrapper[4809]: I1205 12:54:00.886081 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d574f3bf-cb38-4fbb-bb18-0f3135ee2630" path="/var/lib/kubelet/pods/d574f3bf-cb38-4fbb-bb18-0f3135ee2630/volumes" Dec 05 12:54:07 crc kubenswrapper[4809]: I1205 12:54:07.041337 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-nr4kd"] Dec 05 12:54:07 crc kubenswrapper[4809]: I1205 12:54:07.056204 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-nr4kd"] Dec 05 12:54:08 crc kubenswrapper[4809]: I1205 12:54:08.889007 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a08109f6-e55a-4a76-a817-d788341d85b0" path="/var/lib/kubelet/pods/a08109f6-e55a-4a76-a817-d788341d85b0/volumes" Dec 05 12:54:14 crc kubenswrapper[4809]: I1205 12:54:14.046738 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:54:14 crc kubenswrapper[4809]: I1205 12:54:14.047335 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:54:19 crc kubenswrapper[4809]: I1205 12:54:19.293833 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.825143 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9d8657d79-szmgn"] Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.827508 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.831206 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.852183 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9d8657d79-szmgn"] Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.901256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-dns-svc\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.901352 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-sb\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.901448 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvnqq\" (UniqueName: \"kubernetes.io/projected/f29a08fa-0e3d-4417-a96d-631f35511035-kube-api-access-pvnqq\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.901508 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-openstack-cell1\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.901556 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-nb\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:36 crc kubenswrapper[4809]: I1205 12:54:36.901595 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-config\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.004213 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvnqq\" (UniqueName: \"kubernetes.io/projected/f29a08fa-0e3d-4417-a96d-631f35511035-kube-api-access-pvnqq\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.004315 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-openstack-cell1\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.004356 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-nb\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.004401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-config\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.004500 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-dns-svc\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.004553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-sb\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.005208 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-openstack-cell1\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.006201 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-nb\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.006577 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-config\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.006693 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-sb\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.006751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-dns-svc\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.022574 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvnqq\" (UniqueName: \"kubernetes.io/projected/f29a08fa-0e3d-4417-a96d-631f35511035-kube-api-access-pvnqq\") pod \"dnsmasq-dns-9d8657d79-szmgn\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.193267 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:37 crc kubenswrapper[4809]: I1205 12:54:37.796717 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9d8657d79-szmgn"] Dec 05 12:54:37 crc kubenswrapper[4809]: W1205 12:54:37.806323 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf29a08fa_0e3d_4417_a96d_631f35511035.slice/crio-de9fb10f711496b522a984ffb72d9cae1778ceed3779343a4b213f59aea34e75 WatchSource:0}: Error finding container de9fb10f711496b522a984ffb72d9cae1778ceed3779343a4b213f59aea34e75: Status 404 returned error can't find the container with id de9fb10f711496b522a984ffb72d9cae1778ceed3779343a4b213f59aea34e75 Dec 05 12:54:38 crc kubenswrapper[4809]: I1205 12:54:38.431313 4809 generic.go:334] "Generic (PLEG): container finished" podID="f29a08fa-0e3d-4417-a96d-631f35511035" containerID="158b738a8490f99ca4ba74554428c890259db51acdacbc53fbdafacb82c31e89" exitCode=0 Dec 05 12:54:38 crc kubenswrapper[4809]: I1205 12:54:38.431497 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" event={"ID":"f29a08fa-0e3d-4417-a96d-631f35511035","Type":"ContainerDied","Data":"158b738a8490f99ca4ba74554428c890259db51acdacbc53fbdafacb82c31e89"} Dec 05 12:54:38 crc kubenswrapper[4809]: I1205 12:54:38.431831 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" event={"ID":"f29a08fa-0e3d-4417-a96d-631f35511035","Type":"ContainerStarted","Data":"de9fb10f711496b522a984ffb72d9cae1778ceed3779343a4b213f59aea34e75"} Dec 05 12:54:39 crc kubenswrapper[4809]: I1205 12:54:39.444111 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" event={"ID":"f29a08fa-0e3d-4417-a96d-631f35511035","Type":"ContainerStarted","Data":"b1d18b0329b8a086c5f076e9f8141ea14b622a9ca1dcbccdb2f785b2b5d2e76c"} Dec 05 12:54:39 crc kubenswrapper[4809]: I1205 12:54:39.444613 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:39 crc kubenswrapper[4809]: I1205 12:54:39.484405 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" podStartSLOduration=3.484383249 podStartE2EDuration="3.484383249s" podCreationTimestamp="2025-12-05 12:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:54:39.479771044 +0000 UTC m=+6374.870747632" watchObservedRunningTime="2025-12-05 12:54:39.484383249 +0000 UTC m=+6374.875359807" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.047330 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.048491 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.048601 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.050289 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.050428 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" gracePeriod=600 Dec 05 12:54:44 crc kubenswrapper[4809]: E1205 12:54:44.181192 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.273793 4809 scope.go:117] "RemoveContainer" containerID="13b13f2d8be871d2d4e0ba09737ccbfa5805a70eb3c088f9e0111d65fd96ba6d" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.321205 4809 scope.go:117] "RemoveContainer" containerID="9d3bea26c5302a972868cf3ac3d659712c4eaf2bac42f875ceec85c1a391a9be" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.393255 4809 scope.go:117] "RemoveContainer" containerID="949f49a3229508b2640e738584bb8fe8a3049ab23a7427a6f62ddeaae00ccc77" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.512088 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" exitCode=0 Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.512169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391"} Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.512232 4809 scope.go:117] "RemoveContainer" containerID="7e0ee3fefa7029bd441f660ece6169c58fb6143e423ed0039f9d6ad8d0c48340" Dec 05 12:54:44 crc kubenswrapper[4809]: I1205 12:54:44.513566 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:54:44 crc kubenswrapper[4809]: E1205 12:54:44.514259 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.195852 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.279329 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6747d5f4b9-lmbf9"] Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.279609 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" podUID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerName="dnsmasq-dns" containerID="cri-o://868c37abc97d13f339f7a35d70e8a52d14039ff85a0f002a0ba80ba270e35204" gracePeriod=10 Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.450835 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58856dc695-gvqqw"] Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.453022 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.481618 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58856dc695-gvqqw"] Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.566258 4809 generic.go:334] "Generic (PLEG): container finished" podID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerID="868c37abc97d13f339f7a35d70e8a52d14039ff85a0f002a0ba80ba270e35204" exitCode=0 Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.566307 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" event={"ID":"be4ba468-19eb-4918-8fb6-cf9653e6db32","Type":"ContainerDied","Data":"868c37abc97d13f339f7a35d70e8a52d14039ff85a0f002a0ba80ba270e35204"} Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.582975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-ovsdbserver-nb\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.583048 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-openstack-cell1\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.583080 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-config\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.583219 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-ovsdbserver-sb\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.583253 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjflr\" (UniqueName: \"kubernetes.io/projected/07888d82-49b7-4107-ae3c-f34572b10fe0-kube-api-access-gjflr\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.583321 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-dns-svc\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.686139 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-openstack-cell1\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.686489 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-config\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.686678 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-ovsdbserver-sb\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.686706 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjflr\" (UniqueName: \"kubernetes.io/projected/07888d82-49b7-4107-ae3c-f34572b10fe0-kube-api-access-gjflr\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.686785 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-dns-svc\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.686848 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-ovsdbserver-nb\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.688066 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-ovsdbserver-nb\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.688079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-openstack-cell1\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.688768 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-config\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.689010 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-ovsdbserver-sb\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.689436 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07888d82-49b7-4107-ae3c-f34572b10fe0-dns-svc\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.706133 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjflr\" (UniqueName: \"kubernetes.io/projected/07888d82-49b7-4107-ae3c-f34572b10fe0-kube-api-access-gjflr\") pod \"dnsmasq-dns-58856dc695-gvqqw\" (UID: \"07888d82-49b7-4107-ae3c-f34572b10fe0\") " pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.825825 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:47 crc kubenswrapper[4809]: I1205 12:54:47.935929 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.095828 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljlc5\" (UniqueName: \"kubernetes.io/projected/be4ba468-19eb-4918-8fb6-cf9653e6db32-kube-api-access-ljlc5\") pod \"be4ba468-19eb-4918-8fb6-cf9653e6db32\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.096100 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-sb\") pod \"be4ba468-19eb-4918-8fb6-cf9653e6db32\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.096214 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-dns-svc\") pod \"be4ba468-19eb-4918-8fb6-cf9653e6db32\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.096254 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-config\") pod \"be4ba468-19eb-4918-8fb6-cf9653e6db32\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.096359 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-nb\") pod \"be4ba468-19eb-4918-8fb6-cf9653e6db32\" (UID: \"be4ba468-19eb-4918-8fb6-cf9653e6db32\") " Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.100775 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be4ba468-19eb-4918-8fb6-cf9653e6db32-kube-api-access-ljlc5" (OuterVolumeSpecName: "kube-api-access-ljlc5") pod "be4ba468-19eb-4918-8fb6-cf9653e6db32" (UID: "be4ba468-19eb-4918-8fb6-cf9653e6db32"). InnerVolumeSpecName "kube-api-access-ljlc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.168953 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-config" (OuterVolumeSpecName: "config") pod "be4ba468-19eb-4918-8fb6-cf9653e6db32" (UID: "be4ba468-19eb-4918-8fb6-cf9653e6db32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.181263 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "be4ba468-19eb-4918-8fb6-cf9653e6db32" (UID: "be4ba468-19eb-4918-8fb6-cf9653e6db32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.182253 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "be4ba468-19eb-4918-8fb6-cf9653e6db32" (UID: "be4ba468-19eb-4918-8fb6-cf9653e6db32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.193183 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "be4ba468-19eb-4918-8fb6-cf9653e6db32" (UID: "be4ba468-19eb-4918-8fb6-cf9653e6db32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.200149 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.200199 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljlc5\" (UniqueName: \"kubernetes.io/projected/be4ba468-19eb-4918-8fb6-cf9653e6db32-kube-api-access-ljlc5\") on node \"crc\" DevicePath \"\"" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.200214 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.200226 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.200238 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ba468-19eb-4918-8fb6-cf9653e6db32-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.360300 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58856dc695-gvqqw"] Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.577547 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" event={"ID":"07888d82-49b7-4107-ae3c-f34572b10fe0","Type":"ContainerStarted","Data":"ff8185027f214d65cc89bd24bd9ad9112139dabfc193844b8d260cd973c18ced"} Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.580390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" event={"ID":"be4ba468-19eb-4918-8fb6-cf9653e6db32","Type":"ContainerDied","Data":"ee1969a7035f914ccb31f71c6e72620c7e3a245410796956e9b670cf4977ebab"} Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.580423 4809 scope.go:117] "RemoveContainer" containerID="868c37abc97d13f339f7a35d70e8a52d14039ff85a0f002a0ba80ba270e35204" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.580466 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6747d5f4b9-lmbf9" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.649348 4809 scope.go:117] "RemoveContainer" containerID="0ce47b66788d185a474a5dc32dc73ca616112c316062f13a8603608063800aab" Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.687241 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6747d5f4b9-lmbf9"] Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.696062 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6747d5f4b9-lmbf9"] Dec 05 12:54:48 crc kubenswrapper[4809]: I1205 12:54:48.886210 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be4ba468-19eb-4918-8fb6-cf9653e6db32" path="/var/lib/kubelet/pods/be4ba468-19eb-4918-8fb6-cf9653e6db32/volumes" Dec 05 12:54:49 crc kubenswrapper[4809]: I1205 12:54:49.601728 4809 generic.go:334] "Generic (PLEG): container finished" podID="07888d82-49b7-4107-ae3c-f34572b10fe0" containerID="8d27fb3ebf6f72f63ebaa8c6acc089d4e9153aa06609a18dfdd854c5d436b248" exitCode=0 Dec 05 12:54:49 crc kubenswrapper[4809]: I1205 12:54:49.601772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" event={"ID":"07888d82-49b7-4107-ae3c-f34572b10fe0","Type":"ContainerDied","Data":"8d27fb3ebf6f72f63ebaa8c6acc089d4e9153aa06609a18dfdd854c5d436b248"} Dec 05 12:54:50 crc kubenswrapper[4809]: I1205 12:54:50.619069 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" event={"ID":"07888d82-49b7-4107-ae3c-f34572b10fe0","Type":"ContainerStarted","Data":"d20462e22e054e2597ab362f6badfa183e203ed0d863e44394e1c989140bfd35"} Dec 05 12:54:50 crc kubenswrapper[4809]: I1205 12:54:50.619612 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:50 crc kubenswrapper[4809]: I1205 12:54:50.641173 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" podStartSLOduration=3.641156187 podStartE2EDuration="3.641156187s" podCreationTimestamp="2025-12-05 12:54:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 12:54:50.635787063 +0000 UTC m=+6386.026763631" watchObservedRunningTime="2025-12-05 12:54:50.641156187 +0000 UTC m=+6386.032132745" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.691549 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs"] Dec 05 12:54:53 crc kubenswrapper[4809]: E1205 12:54:53.692878 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerName="init" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.692899 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerName="init" Dec 05 12:54:53 crc kubenswrapper[4809]: E1205 12:54:53.692940 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerName="dnsmasq-dns" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.692951 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerName="dnsmasq-dns" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.693362 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="be4ba468-19eb-4918-8fb6-cf9653e6db32" containerName="dnsmasq-dns" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.694767 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.698832 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.698963 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.699143 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.699342 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.709525 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs"] Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.774263 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.774847 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.774973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.775120 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jtcg\" (UniqueName: \"kubernetes.io/projected/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-kube-api-access-5jtcg\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.775254 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.876510 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jtcg\" (UniqueName: \"kubernetes.io/projected/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-kube-api-access-5jtcg\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.876606 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.876674 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.876786 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.876839 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.883596 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.884717 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.888374 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.889209 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:53 crc kubenswrapper[4809]: I1205 12:54:53.893236 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jtcg\" (UniqueName: \"kubernetes.io/projected/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-kube-api-access-5jtcg\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:54 crc kubenswrapper[4809]: I1205 12:54:54.029560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:54:54 crc kubenswrapper[4809]: I1205 12:54:54.688167 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs"] Dec 05 12:54:54 crc kubenswrapper[4809]: W1205 12:54:54.692560 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8d0c6ee_25d0_4e37_aa15_da0204a934d3.slice/crio-a66d7b3eadd3b31346475f1a3adf6876b1d9836316142a42339adab74447fa6a WatchSource:0}: Error finding container a66d7b3eadd3b31346475f1a3adf6876b1d9836316142a42339adab74447fa6a: Status 404 returned error can't find the container with id a66d7b3eadd3b31346475f1a3adf6876b1d9836316142a42339adab74447fa6a Dec 05 12:54:55 crc kubenswrapper[4809]: I1205 12:54:55.708266 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" event={"ID":"b8d0c6ee-25d0-4e37-aa15-da0204a934d3","Type":"ContainerStarted","Data":"a66d7b3eadd3b31346475f1a3adf6876b1d9836316142a42339adab74447fa6a"} Dec 05 12:54:57 crc kubenswrapper[4809]: I1205 12:54:57.827832 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58856dc695-gvqqw" Dec 05 12:54:57 crc kubenswrapper[4809]: I1205 12:54:57.918634 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d8657d79-szmgn"] Dec 05 12:54:57 crc kubenswrapper[4809]: I1205 12:54:57.918865 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" containerName="dnsmasq-dns" containerID="cri-o://b1d18b0329b8a086c5f076e9f8141ea14b622a9ca1dcbccdb2f785b2b5d2e76c" gracePeriod=10 Dec 05 12:54:58 crc kubenswrapper[4809]: I1205 12:54:58.753347 4809 generic.go:334] "Generic (PLEG): container finished" podID="f29a08fa-0e3d-4417-a96d-631f35511035" containerID="b1d18b0329b8a086c5f076e9f8141ea14b622a9ca1dcbccdb2f785b2b5d2e76c" exitCode=0 Dec 05 12:54:58 crc kubenswrapper[4809]: I1205 12:54:58.753675 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" event={"ID":"f29a08fa-0e3d-4417-a96d-631f35511035","Type":"ContainerDied","Data":"b1d18b0329b8a086c5f076e9f8141ea14b622a9ca1dcbccdb2f785b2b5d2e76c"} Dec 05 12:54:58 crc kubenswrapper[4809]: I1205 12:54:58.873263 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:54:58 crc kubenswrapper[4809]: E1205 12:54:58.873882 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:55:02 crc kubenswrapper[4809]: I1205 12:55:02.194188 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.151:5353: connect: connection refused" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.406822 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.445828 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-config\") pod \"f29a08fa-0e3d-4417-a96d-631f35511035\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.445897 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-openstack-cell1\") pod \"f29a08fa-0e3d-4417-a96d-631f35511035\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.446088 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-nb\") pod \"f29a08fa-0e3d-4417-a96d-631f35511035\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.446163 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvnqq\" (UniqueName: \"kubernetes.io/projected/f29a08fa-0e3d-4417-a96d-631f35511035-kube-api-access-pvnqq\") pod \"f29a08fa-0e3d-4417-a96d-631f35511035\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.446189 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-dns-svc\") pod \"f29a08fa-0e3d-4417-a96d-631f35511035\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.446287 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-sb\") pod \"f29a08fa-0e3d-4417-a96d-631f35511035\" (UID: \"f29a08fa-0e3d-4417-a96d-631f35511035\") " Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.462843 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f29a08fa-0e3d-4417-a96d-631f35511035-kube-api-access-pvnqq" (OuterVolumeSpecName: "kube-api-access-pvnqq") pod "f29a08fa-0e3d-4417-a96d-631f35511035" (UID: "f29a08fa-0e3d-4417-a96d-631f35511035"). InnerVolumeSpecName "kube-api-access-pvnqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.509818 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "f29a08fa-0e3d-4417-a96d-631f35511035" (UID: "f29a08fa-0e3d-4417-a96d-631f35511035"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.511238 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f29a08fa-0e3d-4417-a96d-631f35511035" (UID: "f29a08fa-0e3d-4417-a96d-631f35511035"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.512025 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-config" (OuterVolumeSpecName: "config") pod "f29a08fa-0e3d-4417-a96d-631f35511035" (UID: "f29a08fa-0e3d-4417-a96d-631f35511035"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.514374 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f29a08fa-0e3d-4417-a96d-631f35511035" (UID: "f29a08fa-0e3d-4417-a96d-631f35511035"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.528760 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f29a08fa-0e3d-4417-a96d-631f35511035" (UID: "f29a08fa-0e3d-4417-a96d-631f35511035"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.548463 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvnqq\" (UniqueName: \"kubernetes.io/projected/f29a08fa-0e3d-4417-a96d-631f35511035-kube-api-access-pvnqq\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.548495 4809 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.548506 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.548515 4809 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-config\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.548524 4809 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.548532 4809 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f29a08fa-0e3d-4417-a96d-631f35511035-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.817399 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" event={"ID":"b8d0c6ee-25d0-4e37-aa15-da0204a934d3","Type":"ContainerStarted","Data":"9c8f13e9f06d24a3c1eca557b7bb35d52362f0b0394dd981c6cefaa290623542"} Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.821300 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" event={"ID":"f29a08fa-0e3d-4417-a96d-631f35511035","Type":"ContainerDied","Data":"de9fb10f711496b522a984ffb72d9cae1778ceed3779343a4b213f59aea34e75"} Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.821343 4809 scope.go:117] "RemoveContainer" containerID="b1d18b0329b8a086c5f076e9f8141ea14b622a9ca1dcbccdb2f785b2b5d2e76c" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.821418 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d8657d79-szmgn" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.842351 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" podStartSLOduration=2.505778493 podStartE2EDuration="11.842318906s" podCreationTimestamp="2025-12-05 12:54:53 +0000 UTC" firstStartedPulling="2025-12-05 12:54:54.695843751 +0000 UTC m=+6390.086820309" lastFinishedPulling="2025-12-05 12:55:04.032384164 +0000 UTC m=+6399.423360722" observedRunningTime="2025-12-05 12:55:04.837037514 +0000 UTC m=+6400.228014092" watchObservedRunningTime="2025-12-05 12:55:04.842318906 +0000 UTC m=+6400.233295504" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.881402 4809 scope.go:117] "RemoveContainer" containerID="158b738a8490f99ca4ba74554428c890259db51acdacbc53fbdafacb82c31e89" Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.886991 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d8657d79-szmgn"] Dec 05 12:55:04 crc kubenswrapper[4809]: I1205 12:55:04.895381 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9d8657d79-szmgn"] Dec 05 12:55:06 crc kubenswrapper[4809]: I1205 12:55:06.885543 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" path="/var/lib/kubelet/pods/f29a08fa-0e3d-4417-a96d-631f35511035/volumes" Dec 05 12:55:11 crc kubenswrapper[4809]: I1205 12:55:11.873089 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:55:11 crc kubenswrapper[4809]: E1205 12:55:11.874035 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:55:17 crc kubenswrapper[4809]: I1205 12:55:17.959174 4809 generic.go:334] "Generic (PLEG): container finished" podID="b8d0c6ee-25d0-4e37-aa15-da0204a934d3" containerID="9c8f13e9f06d24a3c1eca557b7bb35d52362f0b0394dd981c6cefaa290623542" exitCode=0 Dec 05 12:55:17 crc kubenswrapper[4809]: I1205 12:55:17.959262 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" event={"ID":"b8d0c6ee-25d0-4e37-aa15-da0204a934d3","Type":"ContainerDied","Data":"9c8f13e9f06d24a3c1eca557b7bb35d52362f0b0394dd981c6cefaa290623542"} Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.434885 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.520330 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-pre-adoption-validation-combined-ca-bundle\") pod \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.520480 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jtcg\" (UniqueName: \"kubernetes.io/projected/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-kube-api-access-5jtcg\") pod \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.520572 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ssh-key\") pod \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.520651 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ceph\") pod \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.520720 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-inventory\") pod \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\" (UID: \"b8d0c6ee-25d0-4e37-aa15-da0204a934d3\") " Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.549209 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-kube-api-access-5jtcg" (OuterVolumeSpecName: "kube-api-access-5jtcg") pod "b8d0c6ee-25d0-4e37-aa15-da0204a934d3" (UID: "b8d0c6ee-25d0-4e37-aa15-da0204a934d3"). InnerVolumeSpecName "kube-api-access-5jtcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.549398 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "b8d0c6ee-25d0-4e37-aa15-da0204a934d3" (UID: "b8d0c6ee-25d0-4e37-aa15-da0204a934d3"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.550587 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ceph" (OuterVolumeSpecName: "ceph") pod "b8d0c6ee-25d0-4e37-aa15-da0204a934d3" (UID: "b8d0c6ee-25d0-4e37-aa15-da0204a934d3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.566831 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-inventory" (OuterVolumeSpecName: "inventory") pod "b8d0c6ee-25d0-4e37-aa15-da0204a934d3" (UID: "b8d0c6ee-25d0-4e37-aa15-da0204a934d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.577178 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8d0c6ee-25d0-4e37-aa15-da0204a934d3" (UID: "b8d0c6ee-25d0-4e37-aa15-da0204a934d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.623454 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jtcg\" (UniqueName: \"kubernetes.io/projected/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-kube-api-access-5jtcg\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.623494 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.623509 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.623519 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.623531 4809 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8d0c6ee-25d0-4e37-aa15-da0204a934d3-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.985564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" event={"ID":"b8d0c6ee-25d0-4e37-aa15-da0204a934d3","Type":"ContainerDied","Data":"a66d7b3eadd3b31346475f1a3adf6876b1d9836316142a42339adab74447fa6a"} Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.985602 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a66d7b3eadd3b31346475f1a3adf6876b1d9836316142a42339adab74447fa6a" Dec 05 12:55:19 crc kubenswrapper[4809]: I1205 12:55:19.985703 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs" Dec 05 12:55:25 crc kubenswrapper[4809]: I1205 12:55:25.873244 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:55:25 crc kubenswrapper[4809]: E1205 12:55:25.874394 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.217288 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj"] Dec 05 12:55:31 crc kubenswrapper[4809]: E1205 12:55:31.218032 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" containerName="dnsmasq-dns" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.218045 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" containerName="dnsmasq-dns" Dec 05 12:55:31 crc kubenswrapper[4809]: E1205 12:55:31.218084 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8d0c6ee-25d0-4e37-aa15-da0204a934d3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.218092 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8d0c6ee-25d0-4e37-aa15-da0204a934d3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 05 12:55:31 crc kubenswrapper[4809]: E1205 12:55:31.218106 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" containerName="init" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.218112 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" containerName="init" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.218313 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8d0c6ee-25d0-4e37-aa15-da0204a934d3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.218329 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f29a08fa-0e3d-4417-a96d-631f35511035" containerName="dnsmasq-dns" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.219662 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.225719 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.225771 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.225717 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.227986 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.230729 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj"] Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.399904 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.400271 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sq54\" (UniqueName: \"kubernetes.io/projected/3ee9bb21-400b-4e0a-966c-2625bcc76aae-kube-api-access-9sq54\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.400298 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.400422 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.400493 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.502647 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.502756 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.502797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.502865 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sq54\" (UniqueName: \"kubernetes.io/projected/3ee9bb21-400b-4e0a-966c-2625bcc76aae-kube-api-access-9sq54\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.502891 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.510305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.514433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.514802 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.515451 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.527226 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sq54\" (UniqueName: \"kubernetes.io/projected/3ee9bb21-400b-4e0a-966c-2625bcc76aae-kube-api-access-9sq54\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:31 crc kubenswrapper[4809]: I1205 12:55:31.543026 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 12:55:32 crc kubenswrapper[4809]: I1205 12:55:32.303472 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 12:55:32 crc kubenswrapper[4809]: I1205 12:55:32.321841 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj"] Dec 05 12:55:33 crc kubenswrapper[4809]: I1205 12:55:33.139235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" event={"ID":"3ee9bb21-400b-4e0a-966c-2625bcc76aae","Type":"ContainerStarted","Data":"6af34f6148dde5a114c2f074db702da4b169ee1b5e2f7312207adcc77de3089e"} Dec 05 12:55:34 crc kubenswrapper[4809]: I1205 12:55:34.151978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" event={"ID":"3ee9bb21-400b-4e0a-966c-2625bcc76aae","Type":"ContainerStarted","Data":"100191ce4792fa7497006b921a3bd7b71d0d8a67c0a978dea958f37925194df0"} Dec 05 12:55:34 crc kubenswrapper[4809]: I1205 12:55:34.175932 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" podStartSLOduration=2.5502888280000002 podStartE2EDuration="3.175911743s" podCreationTimestamp="2025-12-05 12:55:31 +0000 UTC" firstStartedPulling="2025-12-05 12:55:32.303210946 +0000 UTC m=+6427.694187504" lastFinishedPulling="2025-12-05 12:55:32.928833861 +0000 UTC m=+6428.319810419" observedRunningTime="2025-12-05 12:55:34.168693159 +0000 UTC m=+6429.559669737" watchObservedRunningTime="2025-12-05 12:55:34.175911743 +0000 UTC m=+6429.566888311" Dec 05 12:55:37 crc kubenswrapper[4809]: I1205 12:55:37.872905 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:55:37 crc kubenswrapper[4809]: E1205 12:55:37.873681 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:55:50 crc kubenswrapper[4809]: I1205 12:55:50.873057 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:55:50 crc kubenswrapper[4809]: E1205 12:55:50.873897 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:56:05 crc kubenswrapper[4809]: I1205 12:56:05.873462 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:56:05 crc kubenswrapper[4809]: E1205 12:56:05.874986 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:56:20 crc kubenswrapper[4809]: I1205 12:56:20.872563 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:56:20 crc kubenswrapper[4809]: E1205 12:56:20.873260 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:56:35 crc kubenswrapper[4809]: I1205 12:56:35.872831 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:56:35 crc kubenswrapper[4809]: E1205 12:56:35.873818 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:56:49 crc kubenswrapper[4809]: I1205 12:56:49.065366 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-ljckw"] Dec 05 12:56:49 crc kubenswrapper[4809]: I1205 12:56:49.076605 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-ljckw"] Dec 05 12:56:50 crc kubenswrapper[4809]: I1205 12:56:50.039811 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-8dfb-account-create-update-kw22n"] Dec 05 12:56:50 crc kubenswrapper[4809]: I1205 12:56:50.054027 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-8dfb-account-create-update-kw22n"] Dec 05 12:56:50 crc kubenswrapper[4809]: I1205 12:56:50.876746 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:56:50 crc kubenswrapper[4809]: E1205 12:56:50.877299 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:56:50 crc kubenswrapper[4809]: I1205 12:56:50.890454 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cd5acc7-eaa1-407a-ac38-c6d904b8e710" path="/var/lib/kubelet/pods/7cd5acc7-eaa1-407a-ac38-c6d904b8e710/volumes" Dec 05 12:56:50 crc kubenswrapper[4809]: I1205 12:56:50.892359 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b825193a-8c7b-45c6-80cd-e929b8771e25" path="/var/lib/kubelet/pods/b825193a-8c7b-45c6-80cd-e929b8771e25/volumes" Dec 05 12:56:59 crc kubenswrapper[4809]: I1205 12:56:59.050924 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-q9566"] Dec 05 12:56:59 crc kubenswrapper[4809]: I1205 12:56:59.067059 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-q9566"] Dec 05 12:57:00 crc kubenswrapper[4809]: I1205 12:57:00.064323 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-2e9e-account-create-update-qh759"] Dec 05 12:57:00 crc kubenswrapper[4809]: I1205 12:57:00.078662 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-2e9e-account-create-update-qh759"] Dec 05 12:57:00 crc kubenswrapper[4809]: I1205 12:57:00.888325 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eba5b4d-1157-4cb8-982d-aca3713a3311" path="/var/lib/kubelet/pods/9eba5b4d-1157-4cb8-982d-aca3713a3311/volumes" Dec 05 12:57:00 crc kubenswrapper[4809]: I1205 12:57:00.889443 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0a6148f-dfac-4671-a743-e16dbf29a4d4" path="/var/lib/kubelet/pods/b0a6148f-dfac-4671-a743-e16dbf29a4d4/volumes" Dec 05 12:57:03 crc kubenswrapper[4809]: I1205 12:57:03.873253 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:57:03 crc kubenswrapper[4809]: E1205 12:57:03.874349 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:57:14 crc kubenswrapper[4809]: I1205 12:57:14.885831 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:57:14 crc kubenswrapper[4809]: E1205 12:57:14.886660 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:57:29 crc kubenswrapper[4809]: I1205 12:57:29.872397 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:57:29 crc kubenswrapper[4809]: E1205 12:57:29.873322 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:57:41 crc kubenswrapper[4809]: I1205 12:57:41.873064 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:57:41 crc kubenswrapper[4809]: E1205 12:57:41.873935 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:57:44 crc kubenswrapper[4809]: I1205 12:57:44.642207 4809 scope.go:117] "RemoveContainer" containerID="b9b3a0f6ee5c8638652650e608983e6164eb5f2c5cf7ea33f2a61b7f7833d350" Dec 05 12:57:44 crc kubenswrapper[4809]: I1205 12:57:44.692524 4809 scope.go:117] "RemoveContainer" containerID="118479dc83bbf3f90a720c0acbf80435a293c0194d6dcb7a9a2271743aac5774" Dec 05 12:57:44 crc kubenswrapper[4809]: I1205 12:57:44.753270 4809 scope.go:117] "RemoveContainer" containerID="fa17c5af786b0d2b9fda3f78e0f50517d6419286baf9ca775ce3d0ee186807d3" Dec 05 12:57:44 crc kubenswrapper[4809]: I1205 12:57:44.824281 4809 scope.go:117] "RemoveContainer" containerID="069cf1f28b90d6d086c50a6997d1575bc89bcbac8de589f870361c9759ded0c3" Dec 05 12:57:48 crc kubenswrapper[4809]: I1205 12:57:48.069115 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-ch655"] Dec 05 12:57:48 crc kubenswrapper[4809]: I1205 12:57:48.081119 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-ch655"] Dec 05 12:57:48 crc kubenswrapper[4809]: I1205 12:57:48.885114 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1230fe8-b1b1-443e-adb6-37287dabebb6" path="/var/lib/kubelet/pods/d1230fe8-b1b1-443e-adb6-37287dabebb6/volumes" Dec 05 12:57:55 crc kubenswrapper[4809]: I1205 12:57:55.873437 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:57:55 crc kubenswrapper[4809]: E1205 12:57:55.874400 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:58:07 crc kubenswrapper[4809]: I1205 12:58:07.873309 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:58:07 crc kubenswrapper[4809]: E1205 12:58:07.874347 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:58:19 crc kubenswrapper[4809]: I1205 12:58:19.872303 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:58:19 crc kubenswrapper[4809]: E1205 12:58:19.873742 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:58:32 crc kubenswrapper[4809]: I1205 12:58:32.872723 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:58:32 crc kubenswrapper[4809]: E1205 12:58:32.873774 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:58:44 crc kubenswrapper[4809]: I1205 12:58:44.962767 4809 scope.go:117] "RemoveContainer" containerID="0c9712a6605652ac0fe2f7207befe9f417e070a32becd9d5062c1f7e86319442" Dec 05 12:58:45 crc kubenswrapper[4809]: I1205 12:58:45.005329 4809 scope.go:117] "RemoveContainer" containerID="b92a0e46f69b0ddf69d1f27dd0e020a6bff979a620ae5e67cb7bddaab66588da" Dec 05 12:58:45 crc kubenswrapper[4809]: I1205 12:58:45.033053 4809 scope.go:117] "RemoveContainer" containerID="2b9b5d181eb96c8a949dbd732200fc47ea41f012754125d0d40bafed33b9adee" Dec 05 12:58:45 crc kubenswrapper[4809]: I1205 12:58:45.075537 4809 scope.go:117] "RemoveContainer" containerID="d53f4578f2ab7748f51f27627f4071ff00be3dc846bf2723ff92133aabce6e5e" Dec 05 12:58:45 crc kubenswrapper[4809]: I1205 12:58:45.121479 4809 scope.go:117] "RemoveContainer" containerID="e8753961dc3469d948ac5d34760c9e083a3055650a0d54ab2419d13aa437d376" Dec 05 12:58:45 crc kubenswrapper[4809]: I1205 12:58:45.151404 4809 scope.go:117] "RemoveContainer" containerID="3f856f8295dd1000ef901ad36705df999653479b2c3afee565a43673c15ac28c" Dec 05 12:58:46 crc kubenswrapper[4809]: I1205 12:58:46.872946 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:58:46 crc kubenswrapper[4809]: E1205 12:58:46.873557 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:58:59 crc kubenswrapper[4809]: I1205 12:58:59.873335 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:58:59 crc kubenswrapper[4809]: E1205 12:58:59.874014 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:59:14 crc kubenswrapper[4809]: I1205 12:59:14.900381 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:59:14 crc kubenswrapper[4809]: E1205 12:59:14.901272 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:59:28 crc kubenswrapper[4809]: I1205 12:59:28.872458 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:59:28 crc kubenswrapper[4809]: E1205 12:59:28.873256 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:59:41 crc kubenswrapper[4809]: I1205 12:59:41.873909 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:59:41 crc kubenswrapper[4809]: E1205 12:59:41.879064 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 12:59:55 crc kubenswrapper[4809]: I1205 12:59:55.872958 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 12:59:56 crc kubenswrapper[4809]: I1205 12:59:56.163660 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"827d4bccfede8bcf3e4988d20dd81c1ec7ceb587b785fd50b3f4bf1f0feb87d9"} Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.153590 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7"] Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.156127 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.158913 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.161825 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.164117 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7"] Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.315853 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f4dbc-fe92-423b-aede-f14c3e321095-secret-volume\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.316223 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlbw7\" (UniqueName: \"kubernetes.io/projected/886f4dbc-fe92-423b-aede-f14c3e321095-kube-api-access-mlbw7\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.316436 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f4dbc-fe92-423b-aede-f14c3e321095-config-volume\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.418868 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f4dbc-fe92-423b-aede-f14c3e321095-config-volume\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.419009 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f4dbc-fe92-423b-aede-f14c3e321095-secret-volume\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.419103 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlbw7\" (UniqueName: \"kubernetes.io/projected/886f4dbc-fe92-423b-aede-f14c3e321095-kube-api-access-mlbw7\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.421274 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f4dbc-fe92-423b-aede-f14c3e321095-config-volume\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.427525 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f4dbc-fe92-423b-aede-f14c3e321095-secret-volume\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.441457 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlbw7\" (UniqueName: \"kubernetes.io/projected/886f4dbc-fe92-423b-aede-f14c3e321095-kube-api-access-mlbw7\") pod \"collect-profiles-29415660-qvzv7\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:00 crc kubenswrapper[4809]: I1205 13:00:00.476336 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:03 crc kubenswrapper[4809]: I1205 13:00:03.088102 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7"] Dec 05 13:00:03 crc kubenswrapper[4809]: I1205 13:00:03.241075 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" event={"ID":"886f4dbc-fe92-423b-aede-f14c3e321095","Type":"ContainerStarted","Data":"4856f963a63077dca3268efa03691bd274ec380b45eecb0bcea35059aa34a510"} Dec 05 13:00:04 crc kubenswrapper[4809]: I1205 13:00:04.252785 4809 generic.go:334] "Generic (PLEG): container finished" podID="886f4dbc-fe92-423b-aede-f14c3e321095" containerID="503b339a61e7c04302fb430ca56f2aa32a7863d49235444049daf6ab7a4d2808" exitCode=0 Dec 05 13:00:04 crc kubenswrapper[4809]: I1205 13:00:04.252882 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" event={"ID":"886f4dbc-fe92-423b-aede-f14c3e321095","Type":"ContainerDied","Data":"503b339a61e7c04302fb430ca56f2aa32a7863d49235444049daf6ab7a4d2808"} Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.620257 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.750139 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f4dbc-fe92-423b-aede-f14c3e321095-config-volume\") pod \"886f4dbc-fe92-423b-aede-f14c3e321095\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.750318 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlbw7\" (UniqueName: \"kubernetes.io/projected/886f4dbc-fe92-423b-aede-f14c3e321095-kube-api-access-mlbw7\") pod \"886f4dbc-fe92-423b-aede-f14c3e321095\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.750386 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f4dbc-fe92-423b-aede-f14c3e321095-secret-volume\") pod \"886f4dbc-fe92-423b-aede-f14c3e321095\" (UID: \"886f4dbc-fe92-423b-aede-f14c3e321095\") " Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.751100 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886f4dbc-fe92-423b-aede-f14c3e321095-config-volume" (OuterVolumeSpecName: "config-volume") pod "886f4dbc-fe92-423b-aede-f14c3e321095" (UID: "886f4dbc-fe92-423b-aede-f14c3e321095"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.761238 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/886f4dbc-fe92-423b-aede-f14c3e321095-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "886f4dbc-fe92-423b-aede-f14c3e321095" (UID: "886f4dbc-fe92-423b-aede-f14c3e321095"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.761325 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/886f4dbc-fe92-423b-aede-f14c3e321095-kube-api-access-mlbw7" (OuterVolumeSpecName: "kube-api-access-mlbw7") pod "886f4dbc-fe92-423b-aede-f14c3e321095" (UID: "886f4dbc-fe92-423b-aede-f14c3e321095"). InnerVolumeSpecName "kube-api-access-mlbw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.853232 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886f4dbc-fe92-423b-aede-f14c3e321095-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.853272 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlbw7\" (UniqueName: \"kubernetes.io/projected/886f4dbc-fe92-423b-aede-f14c3e321095-kube-api-access-mlbw7\") on node \"crc\" DevicePath \"\"" Dec 05 13:00:05 crc kubenswrapper[4809]: I1205 13:00:05.853286 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886f4dbc-fe92-423b-aede-f14c3e321095-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:00:06 crc kubenswrapper[4809]: I1205 13:00:06.274815 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" event={"ID":"886f4dbc-fe92-423b-aede-f14c3e321095","Type":"ContainerDied","Data":"4856f963a63077dca3268efa03691bd274ec380b45eecb0bcea35059aa34a510"} Dec 05 13:00:06 crc kubenswrapper[4809]: I1205 13:00:06.275911 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4856f963a63077dca3268efa03691bd274ec380b45eecb0bcea35059aa34a510" Dec 05 13:00:06 crc kubenswrapper[4809]: I1205 13:00:06.275214 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7" Dec 05 13:00:06 crc kubenswrapper[4809]: I1205 13:00:06.712508 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt"] Dec 05 13:00:06 crc kubenswrapper[4809]: I1205 13:00:06.722108 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415615-6tpwt"] Dec 05 13:00:06 crc kubenswrapper[4809]: I1205 13:00:06.892209 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d40cc40-03ff-48fd-9a15-04b68ce4b435" path="/var/lib/kubelet/pods/1d40cc40-03ff-48fd-9a15-04b68ce4b435/volumes" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.037363 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5qhd2"] Dec 05 13:00:15 crc kubenswrapper[4809]: E1205 13:00:15.038752 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886f4dbc-fe92-423b-aede-f14c3e321095" containerName="collect-profiles" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.038778 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="886f4dbc-fe92-423b-aede-f14c3e321095" containerName="collect-profiles" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.039166 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="886f4dbc-fe92-423b-aede-f14c3e321095" containerName="collect-profiles" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.042695 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.063196 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5qhd2"] Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.159707 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-utilities\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.159800 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n52ls\" (UniqueName: \"kubernetes.io/projected/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-kube-api-access-n52ls\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.160008 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-catalog-content\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.262198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-catalog-content\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.262415 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-utilities\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.262477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n52ls\" (UniqueName: \"kubernetes.io/projected/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-kube-api-access-n52ls\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.262961 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-catalog-content\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.263024 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-utilities\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.280725 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n52ls\" (UniqueName: \"kubernetes.io/projected/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-kube-api-access-n52ls\") pod \"certified-operators-5qhd2\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.386182 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:15 crc kubenswrapper[4809]: I1205 13:00:15.924374 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5qhd2"] Dec 05 13:00:15 crc kubenswrapper[4809]: W1205 13:00:15.932797 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf41e7b85_451c_4e9b_bfad_7c2cbf0e5a62.slice/crio-eaae5a1e1ab3bb50570e588b09aef2c8023c850ae959229f1c7ec92b93e5881c WatchSource:0}: Error finding container eaae5a1e1ab3bb50570e588b09aef2c8023c850ae959229f1c7ec92b93e5881c: Status 404 returned error can't find the container with id eaae5a1e1ab3bb50570e588b09aef2c8023c850ae959229f1c7ec92b93e5881c Dec 05 13:00:16 crc kubenswrapper[4809]: I1205 13:00:16.385707 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qhd2" event={"ID":"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62","Type":"ContainerStarted","Data":"eaae5a1e1ab3bb50570e588b09aef2c8023c850ae959229f1c7ec92b93e5881c"} Dec 05 13:00:16 crc kubenswrapper[4809]: E1205 13:00:16.652533 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf41e7b85_451c_4e9b_bfad_7c2cbf0e5a62.slice/crio-conmon-f3d1276edc01e73cee504f46a01ec32cd980e99bd6645ce89a101662de7aa124.scope\": RecentStats: unable to find data in memory cache]" Dec 05 13:00:17 crc kubenswrapper[4809]: I1205 13:00:17.398398 4809 generic.go:334] "Generic (PLEG): container finished" podID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerID="f3d1276edc01e73cee504f46a01ec32cd980e99bd6645ce89a101662de7aa124" exitCode=0 Dec 05 13:00:17 crc kubenswrapper[4809]: I1205 13:00:17.398504 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qhd2" event={"ID":"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62","Type":"ContainerDied","Data":"f3d1276edc01e73cee504f46a01ec32cd980e99bd6645ce89a101662de7aa124"} Dec 05 13:00:19 crc kubenswrapper[4809]: I1205 13:00:19.432010 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qhd2" event={"ID":"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62","Type":"ContainerStarted","Data":"1e60b9f6e8416a490c70d0829310a1c3bae4f1c1bf55fd7c921a7c7bc1780086"} Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.044106 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-6130-account-create-update-hnvrr"] Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.055132 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-779pj"] Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.066174 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-6130-account-create-update-hnvrr"] Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.074199 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-779pj"] Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.462289 4809 generic.go:334] "Generic (PLEG): container finished" podID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerID="1e60b9f6e8416a490c70d0829310a1c3bae4f1c1bf55fd7c921a7c7bc1780086" exitCode=0 Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.462358 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qhd2" event={"ID":"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62","Type":"ContainerDied","Data":"1e60b9f6e8416a490c70d0829310a1c3bae4f1c1bf55fd7c921a7c7bc1780086"} Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.886454 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a084734-6cf7-42ff-bdc8-21f5ad887112" path="/var/lib/kubelet/pods/3a084734-6cf7-42ff-bdc8-21f5ad887112/volumes" Dec 05 13:00:22 crc kubenswrapper[4809]: I1205 13:00:22.887311 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4adb3f3f-2746-42e1-bd82-6abebbe95479" path="/var/lib/kubelet/pods/4adb3f3f-2746-42e1-bd82-6abebbe95479/volumes" Dec 05 13:00:25 crc kubenswrapper[4809]: I1205 13:00:25.494365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qhd2" event={"ID":"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62","Type":"ContainerStarted","Data":"27888e4148ebbfdd6f39ef3eab9be519a88feda61cb485a7276fea6d9aa1924c"} Dec 05 13:00:25 crc kubenswrapper[4809]: I1205 13:00:25.513855 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5qhd2" podStartSLOduration=3.3543830310000002 podStartE2EDuration="10.513832412s" podCreationTimestamp="2025-12-05 13:00:15 +0000 UTC" firstStartedPulling="2025-12-05 13:00:17.401366861 +0000 UTC m=+6712.792343419" lastFinishedPulling="2025-12-05 13:00:24.560816242 +0000 UTC m=+6719.951792800" observedRunningTime="2025-12-05 13:00:25.512381313 +0000 UTC m=+6720.903357891" watchObservedRunningTime="2025-12-05 13:00:25.513832412 +0000 UTC m=+6720.904808970" Dec 05 13:00:34 crc kubenswrapper[4809]: I1205 13:00:34.040665 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-lgtr2"] Dec 05 13:00:34 crc kubenswrapper[4809]: I1205 13:00:34.056083 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-lgtr2"] Dec 05 13:00:34 crc kubenswrapper[4809]: I1205 13:00:34.884196 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f30dfe-d44a-4316-b5e5-dd5193352cc8" path="/var/lib/kubelet/pods/95f30dfe-d44a-4316-b5e5-dd5193352cc8/volumes" Dec 05 13:00:35 crc kubenswrapper[4809]: I1205 13:00:35.386865 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:35 crc kubenswrapper[4809]: I1205 13:00:35.386903 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:35 crc kubenswrapper[4809]: I1205 13:00:35.436211 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:35 crc kubenswrapper[4809]: I1205 13:00:35.676148 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:35 crc kubenswrapper[4809]: I1205 13:00:35.722777 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5qhd2"] Dec 05 13:00:37 crc kubenswrapper[4809]: I1205 13:00:37.661477 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5qhd2" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="registry-server" containerID="cri-o://27888e4148ebbfdd6f39ef3eab9be519a88feda61cb485a7276fea6d9aa1924c" gracePeriod=2 Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.675986 4809 generic.go:334] "Generic (PLEG): container finished" podID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerID="27888e4148ebbfdd6f39ef3eab9be519a88feda61cb485a7276fea6d9aa1924c" exitCode=0 Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.676472 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qhd2" event={"ID":"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62","Type":"ContainerDied","Data":"27888e4148ebbfdd6f39ef3eab9be519a88feda61cb485a7276fea6d9aa1924c"} Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.676500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qhd2" event={"ID":"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62","Type":"ContainerDied","Data":"eaae5a1e1ab3bb50570e588b09aef2c8023c850ae959229f1c7ec92b93e5881c"} Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.676512 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaae5a1e1ab3bb50570e588b09aef2c8023c850ae959229f1c7ec92b93e5881c" Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.691422 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.788345 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-catalog-content\") pod \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.788533 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-utilities\") pod \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.788715 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n52ls\" (UniqueName: \"kubernetes.io/projected/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-kube-api-access-n52ls\") pod \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\" (UID: \"f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62\") " Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.789369 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-utilities" (OuterVolumeSpecName: "utilities") pod "f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" (UID: "f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.794718 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-kube-api-access-n52ls" (OuterVolumeSpecName: "kube-api-access-n52ls") pod "f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" (UID: "f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62"). InnerVolumeSpecName "kube-api-access-n52ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.841453 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" (UID: "f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.891020 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.891472 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n52ls\" (UniqueName: \"kubernetes.io/projected/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-kube-api-access-n52ls\") on node \"crc\" DevicePath \"\"" Dec 05 13:00:38 crc kubenswrapper[4809]: I1205 13:00:38.891541 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:00:39 crc kubenswrapper[4809]: I1205 13:00:39.692200 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qhd2" Dec 05 13:00:39 crc kubenswrapper[4809]: I1205 13:00:39.750578 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5qhd2"] Dec 05 13:00:39 crc kubenswrapper[4809]: I1205 13:00:39.763796 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5qhd2"] Dec 05 13:00:40 crc kubenswrapper[4809]: I1205 13:00:40.889047 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" path="/var/lib/kubelet/pods/f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62/volumes" Dec 05 13:00:45 crc kubenswrapper[4809]: I1205 13:00:45.292129 4809 scope.go:117] "RemoveContainer" containerID="589c2763c22f289f78f62648b0c5a53b6a47489b8720fd173824cdb148050e16" Dec 05 13:00:45 crc kubenswrapper[4809]: I1205 13:00:45.330034 4809 scope.go:117] "RemoveContainer" containerID="2ac99a5bb22feed1a7518bb3b05d0b84b76c7f8f079c627af9d74f7e11fcfb45" Dec 05 13:00:45 crc kubenswrapper[4809]: I1205 13:00:45.383434 4809 scope.go:117] "RemoveContainer" containerID="f3d033246874cd8481c18d41ea7fd528ad21fdcd51a87ed24f7695a9a9098245" Dec 05 13:00:45 crc kubenswrapper[4809]: I1205 13:00:45.470428 4809 scope.go:117] "RemoveContainer" containerID="e19164098b2dd8317026e51d4f477e6224d7897f83ec09c1044a8f5f6fc5d31c" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.163994 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29415661-mf6gb"] Dec 05 13:01:00 crc kubenswrapper[4809]: E1205 13:01:00.165603 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="extract-utilities" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.165621 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="extract-utilities" Dec 05 13:01:00 crc kubenswrapper[4809]: E1205 13:01:00.165663 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="registry-server" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.165670 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="registry-server" Dec 05 13:01:00 crc kubenswrapper[4809]: E1205 13:01:00.165726 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="extract-content" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.165738 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="extract-content" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.166071 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41e7b85-451c-4e9b-bfad-7c2cbf0e5a62" containerName="registry-server" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.167385 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.177340 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415661-mf6gb"] Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.189326 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt7cz\" (UniqueName: \"kubernetes.io/projected/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-kube-api-access-tt7cz\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.189366 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-combined-ca-bundle\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.189449 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-config-data\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.189723 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-fernet-keys\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.291051 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-config-data\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.291583 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-fernet-keys\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.291792 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt7cz\" (UniqueName: \"kubernetes.io/projected/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-kube-api-access-tt7cz\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.291857 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-combined-ca-bundle\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.297833 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-config-data\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.297939 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-combined-ca-bundle\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.298805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-fernet-keys\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.323868 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt7cz\" (UniqueName: \"kubernetes.io/projected/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-kube-api-access-tt7cz\") pod \"keystone-cron-29415661-mf6gb\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:00 crc kubenswrapper[4809]: I1205 13:01:00.495951 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:01 crc kubenswrapper[4809]: I1205 13:01:01.023339 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415661-mf6gb"] Dec 05 13:01:01 crc kubenswrapper[4809]: W1205 13:01:01.025787 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a1ae2c0_9d01_45d8_b5b4_536a0c76d271.slice/crio-ad1c3ee4804ac43b938f9217503de2144037a7eaad2fbbc4ef8712f08976cc4f WatchSource:0}: Error finding container ad1c3ee4804ac43b938f9217503de2144037a7eaad2fbbc4ef8712f08976cc4f: Status 404 returned error can't find the container with id ad1c3ee4804ac43b938f9217503de2144037a7eaad2fbbc4ef8712f08976cc4f Dec 05 13:01:01 crc kubenswrapper[4809]: I1205 13:01:01.940834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415661-mf6gb" event={"ID":"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271","Type":"ContainerStarted","Data":"3a0bd006b869762c7eda153854d1d8d7bef0c9684b0dc67ec6bcf45542001771"} Dec 05 13:01:01 crc kubenswrapper[4809]: I1205 13:01:01.941104 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415661-mf6gb" event={"ID":"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271","Type":"ContainerStarted","Data":"ad1c3ee4804ac43b938f9217503de2144037a7eaad2fbbc4ef8712f08976cc4f"} Dec 05 13:01:01 crc kubenswrapper[4809]: I1205 13:01:01.966077 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29415661-mf6gb" podStartSLOduration=1.966053297 podStartE2EDuration="1.966053297s" podCreationTimestamp="2025-12-05 13:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 13:01:01.959962693 +0000 UTC m=+6757.350939261" watchObservedRunningTime="2025-12-05 13:01:01.966053297 +0000 UTC m=+6757.357029855" Dec 05 13:01:04 crc kubenswrapper[4809]: I1205 13:01:04.989039 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" containerID="3a0bd006b869762c7eda153854d1d8d7bef0c9684b0dc67ec6bcf45542001771" exitCode=0 Dec 05 13:01:04 crc kubenswrapper[4809]: I1205 13:01:04.989107 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415661-mf6gb" event={"ID":"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271","Type":"ContainerDied","Data":"3a0bd006b869762c7eda153854d1d8d7bef0c9684b0dc67ec6bcf45542001771"} Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.391661 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.530715 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-config-data\") pod \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.530822 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt7cz\" (UniqueName: \"kubernetes.io/projected/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-kube-api-access-tt7cz\") pod \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.530856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-combined-ca-bundle\") pod \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.530912 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-fernet-keys\") pod \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\" (UID: \"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271\") " Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.536196 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-kube-api-access-tt7cz" (OuterVolumeSpecName: "kube-api-access-tt7cz") pod "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" (UID: "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271"). InnerVolumeSpecName "kube-api-access-tt7cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.537405 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" (UID: "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.566704 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" (UID: "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.591174 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-config-data" (OuterVolumeSpecName: "config-data") pod "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" (UID: "8a1ae2c0-9d01-45d8-b5b4-536a0c76d271"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.636705 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt7cz\" (UniqueName: \"kubernetes.io/projected/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-kube-api-access-tt7cz\") on node \"crc\" DevicePath \"\"" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.636757 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.636776 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 13:01:06 crc kubenswrapper[4809]: I1205 13:01:06.636792 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1ae2c0-9d01-45d8-b5b4-536a0c76d271-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 13:01:07 crc kubenswrapper[4809]: I1205 13:01:07.018412 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415661-mf6gb" event={"ID":"8a1ae2c0-9d01-45d8-b5b4-536a0c76d271","Type":"ContainerDied","Data":"ad1c3ee4804ac43b938f9217503de2144037a7eaad2fbbc4ef8712f08976cc4f"} Dec 05 13:01:07 crc kubenswrapper[4809]: I1205 13:01:07.018861 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad1c3ee4804ac43b938f9217503de2144037a7eaad2fbbc4ef8712f08976cc4f" Dec 05 13:01:07 crc kubenswrapper[4809]: I1205 13:01:07.018509 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415661-mf6gb" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.530689 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dw58n"] Dec 05 13:01:21 crc kubenswrapper[4809]: E1205 13:01:21.532046 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" containerName="keystone-cron" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.532069 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" containerName="keystone-cron" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.532521 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a1ae2c0-9d01-45d8-b5b4-536a0c76d271" containerName="keystone-cron" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.535965 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.552694 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dw58n"] Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.686819 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-catalog-content\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.687017 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9pzg\" (UniqueName: \"kubernetes.io/projected/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-kube-api-access-x9pzg\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.687055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-utilities\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.788929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9pzg\" (UniqueName: \"kubernetes.io/projected/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-kube-api-access-x9pzg\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.788995 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-utilities\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.789141 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-catalog-content\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.789759 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-utilities\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.789853 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-catalog-content\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.811170 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9pzg\" (UniqueName: \"kubernetes.io/projected/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-kube-api-access-x9pzg\") pod \"redhat-operators-dw58n\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:21 crc kubenswrapper[4809]: I1205 13:01:21.868198 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:22 crc kubenswrapper[4809]: I1205 13:01:22.365089 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dw58n"] Dec 05 13:01:22 crc kubenswrapper[4809]: W1205 13:01:22.365599 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56eb11b5_5335_4bfc_9de6_a206eefdbfb4.slice/crio-82dfd154d548f1e390e8e922b8c66e79c34a4631f4930db828354a8b835ec43c WatchSource:0}: Error finding container 82dfd154d548f1e390e8e922b8c66e79c34a4631f4930db828354a8b835ec43c: Status 404 returned error can't find the container with id 82dfd154d548f1e390e8e922b8c66e79c34a4631f4930db828354a8b835ec43c Dec 05 13:01:23 crc kubenswrapper[4809]: I1205 13:01:23.201100 4809 generic.go:334] "Generic (PLEG): container finished" podID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerID="72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97" exitCode=0 Dec 05 13:01:23 crc kubenswrapper[4809]: I1205 13:01:23.201201 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dw58n" event={"ID":"56eb11b5-5335-4bfc-9de6-a206eefdbfb4","Type":"ContainerDied","Data":"72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97"} Dec 05 13:01:23 crc kubenswrapper[4809]: I1205 13:01:23.201374 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dw58n" event={"ID":"56eb11b5-5335-4bfc-9de6-a206eefdbfb4","Type":"ContainerStarted","Data":"82dfd154d548f1e390e8e922b8c66e79c34a4631f4930db828354a8b835ec43c"} Dec 05 13:01:23 crc kubenswrapper[4809]: I1205 13:01:23.203511 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:01:24 crc kubenswrapper[4809]: I1205 13:01:24.212452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dw58n" event={"ID":"56eb11b5-5335-4bfc-9de6-a206eefdbfb4","Type":"ContainerStarted","Data":"0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2"} Dec 05 13:01:31 crc kubenswrapper[4809]: I1205 13:01:31.297735 4809 generic.go:334] "Generic (PLEG): container finished" podID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerID="0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2" exitCode=0 Dec 05 13:01:31 crc kubenswrapper[4809]: I1205 13:01:31.297815 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dw58n" event={"ID":"56eb11b5-5335-4bfc-9de6-a206eefdbfb4","Type":"ContainerDied","Data":"0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2"} Dec 05 13:01:32 crc kubenswrapper[4809]: I1205 13:01:32.309681 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dw58n" event={"ID":"56eb11b5-5335-4bfc-9de6-a206eefdbfb4","Type":"ContainerStarted","Data":"b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001"} Dec 05 13:01:32 crc kubenswrapper[4809]: I1205 13:01:32.331921 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dw58n" podStartSLOduration=2.833666682 podStartE2EDuration="11.331902821s" podCreationTimestamp="2025-12-05 13:01:21 +0000 UTC" firstStartedPulling="2025-12-05 13:01:23.203233672 +0000 UTC m=+6778.594210230" lastFinishedPulling="2025-12-05 13:01:31.701469771 +0000 UTC m=+6787.092446369" observedRunningTime="2025-12-05 13:01:32.32777477 +0000 UTC m=+6787.718751348" watchObservedRunningTime="2025-12-05 13:01:32.331902821 +0000 UTC m=+6787.722879379" Dec 05 13:01:41 crc kubenswrapper[4809]: I1205 13:01:41.869074 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:41 crc kubenswrapper[4809]: I1205 13:01:41.869513 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:41 crc kubenswrapper[4809]: I1205 13:01:41.920261 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:42 crc kubenswrapper[4809]: I1205 13:01:42.484908 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:42 crc kubenswrapper[4809]: I1205 13:01:42.544968 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dw58n"] Dec 05 13:01:44 crc kubenswrapper[4809]: I1205 13:01:44.448608 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dw58n" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="registry-server" containerID="cri-o://b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001" gracePeriod=2 Dec 05 13:01:44 crc kubenswrapper[4809]: I1205 13:01:44.991318 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.116443 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9pzg\" (UniqueName: \"kubernetes.io/projected/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-kube-api-access-x9pzg\") pod \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.117083 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-utilities\") pod \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.117380 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-catalog-content\") pod \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\" (UID: \"56eb11b5-5335-4bfc-9de6-a206eefdbfb4\") " Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.118257 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-utilities" (OuterVolumeSpecName: "utilities") pod "56eb11b5-5335-4bfc-9de6-a206eefdbfb4" (UID: "56eb11b5-5335-4bfc-9de6-a206eefdbfb4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.122947 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-kube-api-access-x9pzg" (OuterVolumeSpecName: "kube-api-access-x9pzg") pod "56eb11b5-5335-4bfc-9de6-a206eefdbfb4" (UID: "56eb11b5-5335-4bfc-9de6-a206eefdbfb4"). InnerVolumeSpecName "kube-api-access-x9pzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.219584 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9pzg\" (UniqueName: \"kubernetes.io/projected/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-kube-api-access-x9pzg\") on node \"crc\" DevicePath \"\"" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.219618 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.226296 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56eb11b5-5335-4bfc-9de6-a206eefdbfb4" (UID: "56eb11b5-5335-4bfc-9de6-a206eefdbfb4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.321451 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56eb11b5-5335-4bfc-9de6-a206eefdbfb4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.463320 4809 generic.go:334] "Generic (PLEG): container finished" podID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerID="b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001" exitCode=0 Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.463389 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dw58n" event={"ID":"56eb11b5-5335-4bfc-9de6-a206eefdbfb4","Type":"ContainerDied","Data":"b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001"} Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.463439 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dw58n" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.463473 4809 scope.go:117] "RemoveContainer" containerID="b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.463452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dw58n" event={"ID":"56eb11b5-5335-4bfc-9de6-a206eefdbfb4","Type":"ContainerDied","Data":"82dfd154d548f1e390e8e922b8c66e79c34a4631f4930db828354a8b835ec43c"} Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.499390 4809 scope.go:117] "RemoveContainer" containerID="0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.516158 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dw58n"] Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.526185 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dw58n"] Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.528790 4809 scope.go:117] "RemoveContainer" containerID="72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.601384 4809 scope.go:117] "RemoveContainer" containerID="b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001" Dec 05 13:01:45 crc kubenswrapper[4809]: E1205 13:01:45.601944 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001\": container with ID starting with b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001 not found: ID does not exist" containerID="b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.601979 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001"} err="failed to get container status \"b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001\": rpc error: code = NotFound desc = could not find container \"b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001\": container with ID starting with b2f6065acbf2162d7661ad1146196030aefa1737bbfeff6b9670beee1544f001 not found: ID does not exist" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.602003 4809 scope.go:117] "RemoveContainer" containerID="0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2" Dec 05 13:01:45 crc kubenswrapper[4809]: E1205 13:01:45.602371 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2\": container with ID starting with 0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2 not found: ID does not exist" containerID="0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.602395 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2"} err="failed to get container status \"0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2\": rpc error: code = NotFound desc = could not find container \"0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2\": container with ID starting with 0480659bb2155892e89060f9c8f6091f37dbe85a6176e5ea5ab3c920bbca25b2 not found: ID does not exist" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.602410 4809 scope.go:117] "RemoveContainer" containerID="72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97" Dec 05 13:01:45 crc kubenswrapper[4809]: E1205 13:01:45.602857 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97\": container with ID starting with 72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97 not found: ID does not exist" containerID="72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97" Dec 05 13:01:45 crc kubenswrapper[4809]: I1205 13:01:45.602887 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97"} err="failed to get container status \"72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97\": rpc error: code = NotFound desc = could not find container \"72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97\": container with ID starting with 72f2b7c1bf35f6a78bbf51156acee2dbe3347d368acfecb91aec04a79bf5fa97 not found: ID does not exist" Dec 05 13:01:46 crc kubenswrapper[4809]: I1205 13:01:46.887145 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" path="/var/lib/kubelet/pods/56eb11b5-5335-4bfc-9de6-a206eefdbfb4/volumes" Dec 05 13:02:11 crc kubenswrapper[4809]: I1205 13:02:11.979993 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cnmds"] Dec 05 13:02:11 crc kubenswrapper[4809]: E1205 13:02:11.981127 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="registry-server" Dec 05 13:02:11 crc kubenswrapper[4809]: I1205 13:02:11.981144 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="registry-server" Dec 05 13:02:11 crc kubenswrapper[4809]: E1205 13:02:11.981183 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="extract-utilities" Dec 05 13:02:11 crc kubenswrapper[4809]: I1205 13:02:11.981195 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="extract-utilities" Dec 05 13:02:11 crc kubenswrapper[4809]: E1205 13:02:11.981228 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="extract-content" Dec 05 13:02:11 crc kubenswrapper[4809]: I1205 13:02:11.981237 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="extract-content" Dec 05 13:02:11 crc kubenswrapper[4809]: I1205 13:02:11.981512 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="56eb11b5-5335-4bfc-9de6-a206eefdbfb4" containerName="registry-server" Dec 05 13:02:11 crc kubenswrapper[4809]: I1205 13:02:11.983463 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:11 crc kubenswrapper[4809]: I1205 13:02:11.990111 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cnmds"] Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.084240 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-utilities\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.084610 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-catalog-content\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.084816 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf65c\" (UniqueName: \"kubernetes.io/projected/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-kube-api-access-hf65c\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.174878 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m4swl"] Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.177997 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.186569 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-utilities\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.186903 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-catalog-content\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.187067 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf65c\" (UniqueName: \"kubernetes.io/projected/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-kube-api-access-hf65c\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.187170 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-utilities\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.187473 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-catalog-content\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.211042 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4swl"] Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.216473 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf65c\" (UniqueName: \"kubernetes.io/projected/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-kube-api-access-hf65c\") pod \"community-operators-cnmds\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.289988 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-catalog-content\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.290319 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x2xr\" (UniqueName: \"kubernetes.io/projected/def2417a-496e-4957-b753-8ea225f37fc5-kube-api-access-9x2xr\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.290351 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-utilities\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.316073 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.393024 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-catalog-content\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.393182 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x2xr\" (UniqueName: \"kubernetes.io/projected/def2417a-496e-4957-b753-8ea225f37fc5-kube-api-access-9x2xr\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.393220 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-utilities\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.394029 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-utilities\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.394391 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-catalog-content\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.416277 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x2xr\" (UniqueName: \"kubernetes.io/projected/def2417a-496e-4957-b753-8ea225f37fc5-kube-api-access-9x2xr\") pod \"redhat-marketplace-m4swl\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.503416 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:12 crc kubenswrapper[4809]: I1205 13:02:12.824220 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cnmds"] Dec 05 13:02:13 crc kubenswrapper[4809]: I1205 13:02:13.039969 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4swl"] Dec 05 13:02:13 crc kubenswrapper[4809]: I1205 13:02:13.792623 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerID="dba44b367ff93e4f2f0cf426a8ca8b7427205b978231d0e6e6a9d6ab5b22fe44" exitCode=0 Dec 05 13:02:13 crc kubenswrapper[4809]: I1205 13:02:13.792897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnmds" event={"ID":"8a4fce22-afa9-4d42-9a92-21c6df19e3a4","Type":"ContainerDied","Data":"dba44b367ff93e4f2f0cf426a8ca8b7427205b978231d0e6e6a9d6ab5b22fe44"} Dec 05 13:02:13 crc kubenswrapper[4809]: I1205 13:02:13.793160 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnmds" event={"ID":"8a4fce22-afa9-4d42-9a92-21c6df19e3a4","Type":"ContainerStarted","Data":"6e9792668b7288b770046710abd03297d7f4b435e75c2c0868b88c9d776b2d57"} Dec 05 13:02:13 crc kubenswrapper[4809]: I1205 13:02:13.795919 4809 generic.go:334] "Generic (PLEG): container finished" podID="def2417a-496e-4957-b753-8ea225f37fc5" containerID="665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126" exitCode=0 Dec 05 13:02:13 crc kubenswrapper[4809]: I1205 13:02:13.795992 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4swl" event={"ID":"def2417a-496e-4957-b753-8ea225f37fc5","Type":"ContainerDied","Data":"665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126"} Dec 05 13:02:13 crc kubenswrapper[4809]: I1205 13:02:13.796021 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4swl" event={"ID":"def2417a-496e-4957-b753-8ea225f37fc5","Type":"ContainerStarted","Data":"1dfdc2cd5391d8466eb8eb25fc0fb149e2f0e04e0feae6bab564acaf94300086"} Dec 05 13:02:14 crc kubenswrapper[4809]: I1205 13:02:14.056892 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:02:14 crc kubenswrapper[4809]: I1205 13:02:14.057268 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:02:14 crc kubenswrapper[4809]: I1205 13:02:14.815925 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4swl" event={"ID":"def2417a-496e-4957-b753-8ea225f37fc5","Type":"ContainerStarted","Data":"c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b"} Dec 05 13:02:15 crc kubenswrapper[4809]: I1205 13:02:15.828540 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnmds" event={"ID":"8a4fce22-afa9-4d42-9a92-21c6df19e3a4","Type":"ContainerStarted","Data":"1451ce22e8c7bcc3d2d0c33c608dbf435e663e0f06ed26b0365b8b062b6386d1"} Dec 05 13:02:15 crc kubenswrapper[4809]: I1205 13:02:15.833893 4809 generic.go:334] "Generic (PLEG): container finished" podID="def2417a-496e-4957-b753-8ea225f37fc5" containerID="c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b" exitCode=0 Dec 05 13:02:15 crc kubenswrapper[4809]: I1205 13:02:15.833954 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4swl" event={"ID":"def2417a-496e-4957-b753-8ea225f37fc5","Type":"ContainerDied","Data":"c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b"} Dec 05 13:02:16 crc kubenswrapper[4809]: I1205 13:02:16.851776 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerID="1451ce22e8c7bcc3d2d0c33c608dbf435e663e0f06ed26b0365b8b062b6386d1" exitCode=0 Dec 05 13:02:16 crc kubenswrapper[4809]: I1205 13:02:16.852390 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnmds" event={"ID":"8a4fce22-afa9-4d42-9a92-21c6df19e3a4","Type":"ContainerDied","Data":"1451ce22e8c7bcc3d2d0c33c608dbf435e663e0f06ed26b0365b8b062b6386d1"} Dec 05 13:02:16 crc kubenswrapper[4809]: I1205 13:02:16.855265 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4swl" event={"ID":"def2417a-496e-4957-b753-8ea225f37fc5","Type":"ContainerStarted","Data":"be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae"} Dec 05 13:02:16 crc kubenswrapper[4809]: I1205 13:02:16.914719 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m4swl" podStartSLOduration=2.313489243 podStartE2EDuration="4.914689534s" podCreationTimestamp="2025-12-05 13:02:12 +0000 UTC" firstStartedPulling="2025-12-05 13:02:13.799110805 +0000 UTC m=+6829.190087363" lastFinishedPulling="2025-12-05 13:02:16.400311066 +0000 UTC m=+6831.791287654" observedRunningTime="2025-12-05 13:02:16.896253988 +0000 UTC m=+6832.287230566" watchObservedRunningTime="2025-12-05 13:02:16.914689534 +0000 UTC m=+6832.305666102" Dec 05 13:02:17 crc kubenswrapper[4809]: I1205 13:02:17.868289 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnmds" event={"ID":"8a4fce22-afa9-4d42-9a92-21c6df19e3a4","Type":"ContainerStarted","Data":"43ea9fe0530d1cddaf92f9e43b37b7c9f20ebb635418a43309858e4048e44cfe"} Dec 05 13:02:17 crc kubenswrapper[4809]: I1205 13:02:17.894752 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cnmds" podStartSLOduration=3.445646539 podStartE2EDuration="6.89472705s" podCreationTimestamp="2025-12-05 13:02:11 +0000 UTC" firstStartedPulling="2025-12-05 13:02:13.798023976 +0000 UTC m=+6829.189000544" lastFinishedPulling="2025-12-05 13:02:17.247104497 +0000 UTC m=+6832.638081055" observedRunningTime="2025-12-05 13:02:17.886123288 +0000 UTC m=+6833.277099866" watchObservedRunningTime="2025-12-05 13:02:17.89472705 +0000 UTC m=+6833.285703618" Dec 05 13:02:22 crc kubenswrapper[4809]: I1205 13:02:22.316788 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:22 crc kubenswrapper[4809]: I1205 13:02:22.317529 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:22 crc kubenswrapper[4809]: I1205 13:02:22.366344 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:22 crc kubenswrapper[4809]: I1205 13:02:22.503607 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:22 crc kubenswrapper[4809]: I1205 13:02:22.503690 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:22 crc kubenswrapper[4809]: I1205 13:02:22.570842 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:22 crc kubenswrapper[4809]: I1205 13:02:22.987327 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:23 crc kubenswrapper[4809]: I1205 13:02:23.002925 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.368259 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4swl"] Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.369166 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m4swl" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="registry-server" containerID="cri-o://be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae" gracePeriod=2 Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.567658 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cnmds"] Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.568190 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cnmds" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="registry-server" containerID="cri-o://43ea9fe0530d1cddaf92f9e43b37b7c9f20ebb635418a43309858e4048e44cfe" gracePeriod=2 Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.891289 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.962464 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-catalog-content\") pod \"def2417a-496e-4957-b753-8ea225f37fc5\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.962689 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x2xr\" (UniqueName: \"kubernetes.io/projected/def2417a-496e-4957-b753-8ea225f37fc5-kube-api-access-9x2xr\") pod \"def2417a-496e-4957-b753-8ea225f37fc5\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.962722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-utilities\") pod \"def2417a-496e-4957-b753-8ea225f37fc5\" (UID: \"def2417a-496e-4957-b753-8ea225f37fc5\") " Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.963793 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-utilities" (OuterVolumeSpecName: "utilities") pod "def2417a-496e-4957-b753-8ea225f37fc5" (UID: "def2417a-496e-4957-b753-8ea225f37fc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.964536 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:02:27 crc kubenswrapper[4809]: I1205 13:02:27.978770 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def2417a-496e-4957-b753-8ea225f37fc5-kube-api-access-9x2xr" (OuterVolumeSpecName: "kube-api-access-9x2xr") pod "def2417a-496e-4957-b753-8ea225f37fc5" (UID: "def2417a-496e-4957-b753-8ea225f37fc5"). InnerVolumeSpecName "kube-api-access-9x2xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.002417 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerID="43ea9fe0530d1cddaf92f9e43b37b7c9f20ebb635418a43309858e4048e44cfe" exitCode=0 Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.002564 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnmds" event={"ID":"8a4fce22-afa9-4d42-9a92-21c6df19e3a4","Type":"ContainerDied","Data":"43ea9fe0530d1cddaf92f9e43b37b7c9f20ebb635418a43309858e4048e44cfe"} Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.003275 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "def2417a-496e-4957-b753-8ea225f37fc5" (UID: "def2417a-496e-4957-b753-8ea225f37fc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.005721 4809 generic.go:334] "Generic (PLEG): container finished" podID="def2417a-496e-4957-b753-8ea225f37fc5" containerID="be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae" exitCode=0 Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.005763 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4swl" event={"ID":"def2417a-496e-4957-b753-8ea225f37fc5","Type":"ContainerDied","Data":"be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae"} Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.005797 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4swl" event={"ID":"def2417a-496e-4957-b753-8ea225f37fc5","Type":"ContainerDied","Data":"1dfdc2cd5391d8466eb8eb25fc0fb149e2f0e04e0feae6bab564acaf94300086"} Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.005815 4809 scope.go:117] "RemoveContainer" containerID="be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.005844 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4swl" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.047694 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.048145 4809 scope.go:117] "RemoveContainer" containerID="c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.066528 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/def2417a-496e-4957-b753-8ea225f37fc5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.066576 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x2xr\" (UniqueName: \"kubernetes.io/projected/def2417a-496e-4957-b753-8ea225f37fc5-kube-api-access-9x2xr\") on node \"crc\" DevicePath \"\"" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.071203 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4swl"] Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.084357 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4swl"] Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.094292 4809 scope.go:117] "RemoveContainer" containerID="665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.116154 4809 scope.go:117] "RemoveContainer" containerID="be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae" Dec 05 13:02:28 crc kubenswrapper[4809]: E1205 13:02:28.116622 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae\": container with ID starting with be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae not found: ID does not exist" containerID="be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.116779 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae"} err="failed to get container status \"be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae\": rpc error: code = NotFound desc = could not find container \"be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae\": container with ID starting with be40db2da6b5156f0fb1b9463dba73a1b82429868fd51054e0c0cf12ae84eaae not found: ID does not exist" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.116817 4809 scope.go:117] "RemoveContainer" containerID="c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b" Dec 05 13:02:28 crc kubenswrapper[4809]: E1205 13:02:28.117190 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b\": container with ID starting with c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b not found: ID does not exist" containerID="c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.117225 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b"} err="failed to get container status \"c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b\": rpc error: code = NotFound desc = could not find container \"c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b\": container with ID starting with c9c14e6fec1809e411a18dcf262b9ab656704ea021578daefa32dfa6a01a557b not found: ID does not exist" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.117246 4809 scope.go:117] "RemoveContainer" containerID="665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126" Dec 05 13:02:28 crc kubenswrapper[4809]: E1205 13:02:28.117459 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126\": container with ID starting with 665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126 not found: ID does not exist" containerID="665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.117483 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126"} err="failed to get container status \"665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126\": rpc error: code = NotFound desc = could not find container \"665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126\": container with ID starting with 665e920001f629676219d686a1b09cf7f523393d14f243203a7ee11bafa23126 not found: ID does not exist" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.168381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-utilities\") pod \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.168588 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf65c\" (UniqueName: \"kubernetes.io/projected/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-kube-api-access-hf65c\") pod \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.168709 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-catalog-content\") pod \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\" (UID: \"8a4fce22-afa9-4d42-9a92-21c6df19e3a4\") " Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.169285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-utilities" (OuterVolumeSpecName: "utilities") pod "8a4fce22-afa9-4d42-9a92-21c6df19e3a4" (UID: "8a4fce22-afa9-4d42-9a92-21c6df19e3a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.175193 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-kube-api-access-hf65c" (OuterVolumeSpecName: "kube-api-access-hf65c") pod "8a4fce22-afa9-4d42-9a92-21c6df19e3a4" (UID: "8a4fce22-afa9-4d42-9a92-21c6df19e3a4"). InnerVolumeSpecName "kube-api-access-hf65c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.227747 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a4fce22-afa9-4d42-9a92-21c6df19e3a4" (UID: "8a4fce22-afa9-4d42-9a92-21c6df19e3a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.271771 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf65c\" (UniqueName: \"kubernetes.io/projected/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-kube-api-access-hf65c\") on node \"crc\" DevicePath \"\"" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.271817 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.271827 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4fce22-afa9-4d42-9a92-21c6df19e3a4-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:02:28 crc kubenswrapper[4809]: I1205 13:02:28.884828 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="def2417a-496e-4957-b753-8ea225f37fc5" path="/var/lib/kubelet/pods/def2417a-496e-4957-b753-8ea225f37fc5/volumes" Dec 05 13:02:29 crc kubenswrapper[4809]: I1205 13:02:29.017571 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnmds" event={"ID":"8a4fce22-afa9-4d42-9a92-21c6df19e3a4","Type":"ContainerDied","Data":"6e9792668b7288b770046710abd03297d7f4b435e75c2c0868b88c9d776b2d57"} Dec 05 13:02:29 crc kubenswrapper[4809]: I1205 13:02:29.017672 4809 scope.go:117] "RemoveContainer" containerID="43ea9fe0530d1cddaf92f9e43b37b7c9f20ebb635418a43309858e4048e44cfe" Dec 05 13:02:29 crc kubenswrapper[4809]: I1205 13:02:29.017589 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnmds" Dec 05 13:02:29 crc kubenswrapper[4809]: I1205 13:02:29.043070 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cnmds"] Dec 05 13:02:29 crc kubenswrapper[4809]: I1205 13:02:29.050137 4809 scope.go:117] "RemoveContainer" containerID="1451ce22e8c7bcc3d2d0c33c608dbf435e663e0f06ed26b0365b8b062b6386d1" Dec 05 13:02:29 crc kubenswrapper[4809]: I1205 13:02:29.052175 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cnmds"] Dec 05 13:02:29 crc kubenswrapper[4809]: I1205 13:02:29.073316 4809 scope.go:117] "RemoveContainer" containerID="dba44b367ff93e4f2f0cf426a8ca8b7427205b978231d0e6e6a9d6ab5b22fe44" Dec 05 13:02:30 crc kubenswrapper[4809]: I1205 13:02:30.889146 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" path="/var/lib/kubelet/pods/8a4fce22-afa9-4d42-9a92-21c6df19e3a4/volumes" Dec 05 13:02:43 crc kubenswrapper[4809]: I1205 13:02:43.071189 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-pkhwb"] Dec 05 13:02:43 crc kubenswrapper[4809]: I1205 13:02:43.084526 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-8233-account-create-update-rwtqt"] Dec 05 13:02:43 crc kubenswrapper[4809]: I1205 13:02:43.097511 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-8233-account-create-update-rwtqt"] Dec 05 13:02:43 crc kubenswrapper[4809]: I1205 13:02:43.136345 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-pkhwb"] Dec 05 13:02:44 crc kubenswrapper[4809]: I1205 13:02:44.047106 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:02:44 crc kubenswrapper[4809]: I1205 13:02:44.047178 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:02:44 crc kubenswrapper[4809]: I1205 13:02:44.886181 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f2ff1bb-c9f5-4772-814c-71ed1d78ae43" path="/var/lib/kubelet/pods/1f2ff1bb-c9f5-4772-814c-71ed1d78ae43/volumes" Dec 05 13:02:44 crc kubenswrapper[4809]: I1205 13:02:44.887241 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3563aa06-65da-4383-ab21-39c19d9ab0c3" path="/var/lib/kubelet/pods/3563aa06-65da-4383-ab21-39c19d9ab0c3/volumes" Dec 05 13:02:45 crc kubenswrapper[4809]: I1205 13:02:45.652952 4809 scope.go:117] "RemoveContainer" containerID="8d5c13451a3cd9cda653ba05d893de8041bb378443aae68e46fc8ae76fd82eb9" Dec 05 13:02:45 crc kubenswrapper[4809]: I1205 13:02:45.686047 4809 scope.go:117] "RemoveContainer" containerID="64ed934b366e0c56e6de906ae92943b3f634dc94089bb7eaad5c0940d8603ab3" Dec 05 13:02:54 crc kubenswrapper[4809]: I1205 13:02:54.061863 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-5sxdg"] Dec 05 13:02:54 crc kubenswrapper[4809]: I1205 13:02:54.076308 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-5sxdg"] Dec 05 13:02:54 crc kubenswrapper[4809]: I1205 13:02:54.888859 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6018b72-920a-4d96-8351-45e1ee9810f6" path="/var/lib/kubelet/pods/b6018b72-920a-4d96-8351-45e1ee9810f6/volumes" Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.047179 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.047709 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.047761 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.048607 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"827d4bccfede8bcf3e4988d20dd81c1ec7ceb587b785fd50b3f4bf1f0feb87d9"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.048685 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://827d4bccfede8bcf3e4988d20dd81c1ec7ceb587b785fd50b3f4bf1f0feb87d9" gracePeriod=600 Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.536274 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="827d4bccfede8bcf3e4988d20dd81c1ec7ceb587b785fd50b3f4bf1f0feb87d9" exitCode=0 Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.536344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"827d4bccfede8bcf3e4988d20dd81c1ec7ceb587b785fd50b3f4bf1f0feb87d9"} Dec 05 13:03:14 crc kubenswrapper[4809]: I1205 13:03:14.536614 4809 scope.go:117] "RemoveContainer" containerID="c5cd13afc454d34090810e59e94a8d53e03f35ba7aef35de0267fec591f5f391" Dec 05 13:03:15 crc kubenswrapper[4809]: I1205 13:03:15.547855 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80"} Dec 05 13:03:17 crc kubenswrapper[4809]: I1205 13:03:17.039919 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-tnb4v"] Dec 05 13:03:17 crc kubenswrapper[4809]: I1205 13:03:17.052116 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-e20e-account-create-update-f4ght"] Dec 05 13:03:17 crc kubenswrapper[4809]: I1205 13:03:17.060594 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-e20e-account-create-update-f4ght"] Dec 05 13:03:17 crc kubenswrapper[4809]: I1205 13:03:17.069295 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-tnb4v"] Dec 05 13:03:18 crc kubenswrapper[4809]: I1205 13:03:18.889467 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf396280-ba54-401d-b6ac-6aa8df85524f" path="/var/lib/kubelet/pods/bf396280-ba54-401d-b6ac-6aa8df85524f/volumes" Dec 05 13:03:18 crc kubenswrapper[4809]: I1205 13:03:18.890393 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f79a7f59-b085-4378-8e8f-b28914905c8e" path="/var/lib/kubelet/pods/f79a7f59-b085-4378-8e8f-b28914905c8e/volumes" Dec 05 13:03:29 crc kubenswrapper[4809]: I1205 13:03:29.063484 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-r7h9j"] Dec 05 13:03:29 crc kubenswrapper[4809]: I1205 13:03:29.075388 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-r7h9j"] Dec 05 13:03:30 crc kubenswrapper[4809]: I1205 13:03:30.889202 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4db0f5c6-097d-4923-be05-a74f8ddb5fbe" path="/var/lib/kubelet/pods/4db0f5c6-097d-4923-be05-a74f8ddb5fbe/volumes" Dec 05 13:03:45 crc kubenswrapper[4809]: I1205 13:03:45.864335 4809 scope.go:117] "RemoveContainer" containerID="4be7f636c4c494a709b8c3fa18951559d0d867fbc0e6453d63cfb10c93df2c52" Dec 05 13:03:45 crc kubenswrapper[4809]: I1205 13:03:45.937127 4809 scope.go:117] "RemoveContainer" containerID="2cc76cd0c7d3848fa14b2566aa5d299351979a235ff9f92b1b232dd2131f325b" Dec 05 13:03:45 crc kubenswrapper[4809]: I1205 13:03:45.995134 4809 scope.go:117] "RemoveContainer" containerID="1e5d63d0c6bd6af6d27cab8a650b614c2e112c0ed70a541c0b81fa8e503eabaf" Dec 05 13:03:46 crc kubenswrapper[4809]: I1205 13:03:46.024795 4809 scope.go:117] "RemoveContainer" containerID="cde4ef13a0db116a2b4cd335147e623b31b5889a518bf06b98f251f29002d7f4" Dec 05 13:05:44 crc kubenswrapper[4809]: I1205 13:05:44.046552 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:05:44 crc kubenswrapper[4809]: I1205 13:05:44.047282 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:06:09 crc kubenswrapper[4809]: I1205 13:06:09.373042 4809 generic.go:334] "Generic (PLEG): container finished" podID="3ee9bb21-400b-4e0a-966c-2625bcc76aae" containerID="100191ce4792fa7497006b921a3bd7b71d0d8a67c0a978dea958f37925194df0" exitCode=0 Dec 05 13:06:09 crc kubenswrapper[4809]: I1205 13:06:09.373671 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" event={"ID":"3ee9bb21-400b-4e0a-966c-2625bcc76aae","Type":"ContainerDied","Data":"100191ce4792fa7497006b921a3bd7b71d0d8a67c0a978dea958f37925194df0"} Dec 05 13:06:10 crc kubenswrapper[4809]: I1205 13:06:10.956411 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.080170 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ceph\") pod \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.080474 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-inventory\") pod \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.082358 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-tripleo-cleanup-combined-ca-bundle\") pod \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.082467 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sq54\" (UniqueName: \"kubernetes.io/projected/3ee9bb21-400b-4e0a-966c-2625bcc76aae-kube-api-access-9sq54\") pod \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.082513 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ssh-key\") pod \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\" (UID: \"3ee9bb21-400b-4e0a-966c-2625bcc76aae\") " Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.089798 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "3ee9bb21-400b-4e0a-966c-2625bcc76aae" (UID: "3ee9bb21-400b-4e0a-966c-2625bcc76aae"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.091754 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ceph" (OuterVolumeSpecName: "ceph") pod "3ee9bb21-400b-4e0a-966c-2625bcc76aae" (UID: "3ee9bb21-400b-4e0a-966c-2625bcc76aae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.093458 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee9bb21-400b-4e0a-966c-2625bcc76aae-kube-api-access-9sq54" (OuterVolumeSpecName: "kube-api-access-9sq54") pod "3ee9bb21-400b-4e0a-966c-2625bcc76aae" (UID: "3ee9bb21-400b-4e0a-966c-2625bcc76aae"). InnerVolumeSpecName "kube-api-access-9sq54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.121552 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-inventory" (OuterVolumeSpecName: "inventory") pod "3ee9bb21-400b-4e0a-966c-2625bcc76aae" (UID: "3ee9bb21-400b-4e0a-966c-2625bcc76aae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.150853 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3ee9bb21-400b-4e0a-966c-2625bcc76aae" (UID: "3ee9bb21-400b-4e0a-966c-2625bcc76aae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.186010 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.186058 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.186073 4809 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.186085 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sq54\" (UniqueName: \"kubernetes.io/projected/3ee9bb21-400b-4e0a-966c-2625bcc76aae-kube-api-access-9sq54\") on node \"crc\" DevicePath \"\"" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.186094 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ee9bb21-400b-4e0a-966c-2625bcc76aae-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.402450 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" event={"ID":"3ee9bb21-400b-4e0a-966c-2625bcc76aae","Type":"ContainerDied","Data":"6af34f6148dde5a114c2f074db702da4b169ee1b5e2f7312207adcc77de3089e"} Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.402492 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6af34f6148dde5a114c2f074db702da4b169ee1b5e2f7312207adcc77de3089e" Dec 05 13:06:11 crc kubenswrapper[4809]: I1205 13:06:11.402764 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj" Dec 05 13:06:14 crc kubenswrapper[4809]: I1205 13:06:14.046230 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:06:14 crc kubenswrapper[4809]: I1205 13:06:14.046936 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.726228 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rjcr7"] Dec 05 13:06:19 crc kubenswrapper[4809]: E1205 13:06:19.727912 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="registry-server" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.727933 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="registry-server" Dec 05 13:06:19 crc kubenswrapper[4809]: E1205 13:06:19.727969 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="extract-content" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.727978 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="extract-content" Dec 05 13:06:19 crc kubenswrapper[4809]: E1205 13:06:19.728012 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="extract-utilities" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.728019 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="extract-utilities" Dec 05 13:06:19 crc kubenswrapper[4809]: E1205 13:06:19.728573 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="extract-utilities" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.728586 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="extract-utilities" Dec 05 13:06:19 crc kubenswrapper[4809]: E1205 13:06:19.728601 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="registry-server" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.728608 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="registry-server" Dec 05 13:06:19 crc kubenswrapper[4809]: E1205 13:06:19.728621 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee9bb21-400b-4e0a-966c-2625bcc76aae" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.728653 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee9bb21-400b-4e0a-966c-2625bcc76aae" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 05 13:06:19 crc kubenswrapper[4809]: E1205 13:06:19.728684 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="extract-content" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.728695 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="extract-content" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.729027 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee9bb21-400b-4e0a-966c-2625bcc76aae" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.729047 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a4fce22-afa9-4d42-9a92-21c6df19e3a4" containerName="registry-server" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.729072 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="def2417a-496e-4957-b753-8ea225f37fc5" containerName="registry-server" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.730885 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.734383 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.734481 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.734654 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.734627 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.737702 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rjcr7"] Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.775746 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-inventory\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.776005 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.776135 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.776210 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ceph\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.776368 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb2zd\" (UniqueName: \"kubernetes.io/projected/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-kube-api-access-mb2zd\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.877931 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb2zd\" (UniqueName: \"kubernetes.io/projected/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-kube-api-access-mb2zd\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.878017 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-inventory\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.878115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.878233 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.878314 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ceph\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.884412 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.884815 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-inventory\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.894306 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ceph\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.895030 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:19 crc kubenswrapper[4809]: I1205 13:06:19.895727 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb2zd\" (UniqueName: \"kubernetes.io/projected/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-kube-api-access-mb2zd\") pod \"bootstrap-openstack-openstack-cell1-rjcr7\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:20 crc kubenswrapper[4809]: I1205 13:06:20.065764 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:06:20 crc kubenswrapper[4809]: I1205 13:06:20.596588 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rjcr7"] Dec 05 13:06:21 crc kubenswrapper[4809]: I1205 13:06:21.513132 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" event={"ID":"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca","Type":"ContainerStarted","Data":"9ccd660f820bd107bad8b4d91e4375eecd6a6e93ce4e6c1a5bfe94ee78a6cb56"} Dec 05 13:06:22 crc kubenswrapper[4809]: I1205 13:06:22.523565 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" event={"ID":"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca","Type":"ContainerStarted","Data":"66e9eb389c985708d92fd8287b62bf1cc05bfce3313f53fc0b8a63b8a3ace783"} Dec 05 13:06:22 crc kubenswrapper[4809]: I1205 13:06:22.549314 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" podStartSLOduration=2.557306626 podStartE2EDuration="3.549291254s" podCreationTimestamp="2025-12-05 13:06:19 +0000 UTC" firstStartedPulling="2025-12-05 13:06:20.603012122 +0000 UTC m=+7075.993988680" lastFinishedPulling="2025-12-05 13:06:21.59499675 +0000 UTC m=+7076.985973308" observedRunningTime="2025-12-05 13:06:22.537132147 +0000 UTC m=+7077.928108725" watchObservedRunningTime="2025-12-05 13:06:22.549291254 +0000 UTC m=+7077.940267812" Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.047115 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.047721 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.047770 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.048528 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.048594 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" gracePeriod=600 Dec 05 13:06:44 crc kubenswrapper[4809]: E1205 13:06:44.700957 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.757191 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" exitCode=0 Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.757246 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80"} Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.757302 4809 scope.go:117] "RemoveContainer" containerID="827d4bccfede8bcf3e4988d20dd81c1ec7ceb587b785fd50b3f4bf1f0feb87d9" Dec 05 13:06:44 crc kubenswrapper[4809]: I1205 13:06:44.758292 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:06:44 crc kubenswrapper[4809]: E1205 13:06:44.758591 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:06:46 crc kubenswrapper[4809]: I1205 13:06:46.227286 4809 scope.go:117] "RemoveContainer" containerID="27888e4148ebbfdd6f39ef3eab9be519a88feda61cb485a7276fea6d9aa1924c" Dec 05 13:06:46 crc kubenswrapper[4809]: I1205 13:06:46.249692 4809 scope.go:117] "RemoveContainer" containerID="f3d1276edc01e73cee504f46a01ec32cd980e99bd6645ce89a101662de7aa124" Dec 05 13:06:46 crc kubenswrapper[4809]: I1205 13:06:46.271958 4809 scope.go:117] "RemoveContainer" containerID="1e60b9f6e8416a490c70d0829310a1c3bae4f1c1bf55fd7c921a7c7bc1780086" Dec 05 13:06:55 crc kubenswrapper[4809]: I1205 13:06:55.872320 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:06:55 crc kubenswrapper[4809]: E1205 13:06:55.873463 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:07:08 crc kubenswrapper[4809]: I1205 13:07:08.872914 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:07:08 crc kubenswrapper[4809]: E1205 13:07:08.873802 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:07:21 crc kubenswrapper[4809]: I1205 13:07:21.872903 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:07:21 crc kubenswrapper[4809]: E1205 13:07:21.873763 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:07:36 crc kubenswrapper[4809]: I1205 13:07:36.873014 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:07:36 crc kubenswrapper[4809]: E1205 13:07:36.873751 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:07:50 crc kubenswrapper[4809]: I1205 13:07:50.879634 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:07:50 crc kubenswrapper[4809]: E1205 13:07:50.880299 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:08:01 crc kubenswrapper[4809]: I1205 13:08:01.872374 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:08:01 crc kubenswrapper[4809]: E1205 13:08:01.873254 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:08:16 crc kubenswrapper[4809]: I1205 13:08:16.872800 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:08:16 crc kubenswrapper[4809]: E1205 13:08:16.873621 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:08:30 crc kubenswrapper[4809]: I1205 13:08:30.873202 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:08:30 crc kubenswrapper[4809]: E1205 13:08:30.874265 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:08:43 crc kubenswrapper[4809]: I1205 13:08:43.873412 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:08:43 crc kubenswrapper[4809]: E1205 13:08:43.875944 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:08:57 crc kubenswrapper[4809]: I1205 13:08:57.874417 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:08:57 crc kubenswrapper[4809]: E1205 13:08:57.875048 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:09:11 crc kubenswrapper[4809]: I1205 13:09:11.873296 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:09:11 crc kubenswrapper[4809]: E1205 13:09:11.874901 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:09:23 crc kubenswrapper[4809]: I1205 13:09:23.872439 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:09:23 crc kubenswrapper[4809]: E1205 13:09:23.873178 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:09:27 crc kubenswrapper[4809]: I1205 13:09:27.426508 4809 generic.go:334] "Generic (PLEG): container finished" podID="49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" containerID="66e9eb389c985708d92fd8287b62bf1cc05bfce3313f53fc0b8a63b8a3ace783" exitCode=0 Dec 05 13:09:27 crc kubenswrapper[4809]: I1205 13:09:27.426779 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" event={"ID":"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca","Type":"ContainerDied","Data":"66e9eb389c985708d92fd8287b62bf1cc05bfce3313f53fc0b8a63b8a3ace783"} Dec 05 13:09:28 crc kubenswrapper[4809]: I1205 13:09:28.888892 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.041108 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ceph\") pod \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.041472 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb2zd\" (UniqueName: \"kubernetes.io/projected/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-kube-api-access-mb2zd\") pod \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.041552 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ssh-key\") pod \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.041580 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-bootstrap-combined-ca-bundle\") pod \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.041732 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-inventory\") pod \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\" (UID: \"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca\") " Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.046771 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-kube-api-access-mb2zd" (OuterVolumeSpecName: "kube-api-access-mb2zd") pod "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" (UID: "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca"). InnerVolumeSpecName "kube-api-access-mb2zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.047720 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ceph" (OuterVolumeSpecName: "ceph") pod "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" (UID: "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.054999 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" (UID: "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.073802 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-inventory" (OuterVolumeSpecName: "inventory") pod "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" (UID: "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.081438 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" (UID: "49d66a02-1ee1-4e87-b1f2-d7fbcd340eca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.144812 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.144868 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb2zd\" (UniqueName: \"kubernetes.io/projected/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-kube-api-access-mb2zd\") on node \"crc\" DevicePath \"\"" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.144887 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.144899 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.144914 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49d66a02-1ee1-4e87-b1f2-d7fbcd340eca-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.449437 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" event={"ID":"49d66a02-1ee1-4e87-b1f2-d7fbcd340eca","Type":"ContainerDied","Data":"9ccd660f820bd107bad8b4d91e4375eecd6a6e93ce4e6c1a5bfe94ee78a6cb56"} Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.449491 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ccd660f820bd107bad8b4d91e4375eecd6a6e93ce4e6c1a5bfe94ee78a6cb56" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.449453 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rjcr7" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.547596 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-99s2k"] Dec 05 13:09:29 crc kubenswrapper[4809]: E1205 13:09:29.548105 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" containerName="bootstrap-openstack-openstack-cell1" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.548124 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" containerName="bootstrap-openstack-openstack-cell1" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.548359 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d66a02-1ee1-4e87-b1f2-d7fbcd340eca" containerName="bootstrap-openstack-openstack-cell1" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.549169 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.555568 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.555581 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.555739 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.557318 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.558439 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-99s2k"] Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.654612 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-inventory\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.654695 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ssh-key\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.654728 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ceph\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.654790 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk2jz\" (UniqueName: \"kubernetes.io/projected/2ba9f48f-f049-46da-b82b-62785eff9fd3-kube-api-access-kk2jz\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.756557 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-inventory\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.756670 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ssh-key\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.756717 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ceph\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.756809 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk2jz\" (UniqueName: \"kubernetes.io/projected/2ba9f48f-f049-46da-b82b-62785eff9fd3-kube-api-access-kk2jz\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.761773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-inventory\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.761821 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ceph\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.762580 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ssh-key\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.773295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk2jz\" (UniqueName: \"kubernetes.io/projected/2ba9f48f-f049-46da-b82b-62785eff9fd3-kube-api-access-kk2jz\") pod \"download-cache-openstack-openstack-cell1-99s2k\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:29 crc kubenswrapper[4809]: I1205 13:09:29.874829 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:09:30 crc kubenswrapper[4809]: I1205 13:09:30.432558 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-99s2k"] Dec 05 13:09:30 crc kubenswrapper[4809]: W1205 13:09:30.443186 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ba9f48f_f049_46da_b82b_62785eff9fd3.slice/crio-7a669fa03f221da58bea475cdda040bd8f5b3d5da4751f7f5da7b9297bfe85c6 WatchSource:0}: Error finding container 7a669fa03f221da58bea475cdda040bd8f5b3d5da4751f7f5da7b9297bfe85c6: Status 404 returned error can't find the container with id 7a669fa03f221da58bea475cdda040bd8f5b3d5da4751f7f5da7b9297bfe85c6 Dec 05 13:09:30 crc kubenswrapper[4809]: I1205 13:09:30.445557 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:09:30 crc kubenswrapper[4809]: I1205 13:09:30.461448 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" event={"ID":"2ba9f48f-f049-46da-b82b-62785eff9fd3","Type":"ContainerStarted","Data":"7a669fa03f221da58bea475cdda040bd8f5b3d5da4751f7f5da7b9297bfe85c6"} Dec 05 13:09:31 crc kubenswrapper[4809]: I1205 13:09:31.473500 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" event={"ID":"2ba9f48f-f049-46da-b82b-62785eff9fd3","Type":"ContainerStarted","Data":"b3ca428f1edf967449e79e1cd8c1defad4e0ea93b24d84655e90ef6af9c8fb38"} Dec 05 13:09:31 crc kubenswrapper[4809]: I1205 13:09:31.501321 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" podStartSLOduration=1.995818076 podStartE2EDuration="2.501300872s" podCreationTimestamp="2025-12-05 13:09:29 +0000 UTC" firstStartedPulling="2025-12-05 13:09:30.445298256 +0000 UTC m=+7265.836274814" lastFinishedPulling="2025-12-05 13:09:30.950781052 +0000 UTC m=+7266.341757610" observedRunningTime="2025-12-05 13:09:31.49156301 +0000 UTC m=+7266.882539568" watchObservedRunningTime="2025-12-05 13:09:31.501300872 +0000 UTC m=+7266.892277430" Dec 05 13:09:36 crc kubenswrapper[4809]: I1205 13:09:36.872477 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:09:36 crc kubenswrapper[4809]: E1205 13:09:36.873266 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:09:50 crc kubenswrapper[4809]: I1205 13:09:50.873048 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:09:50 crc kubenswrapper[4809]: E1205 13:09:50.873859 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:10:05 crc kubenswrapper[4809]: I1205 13:10:05.872179 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:10:05 crc kubenswrapper[4809]: E1205 13:10:05.872992 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.521548 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mrp9n"] Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.524819 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.540518 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrp9n"] Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.627836 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-utilities\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.628033 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99twn\" (UniqueName: \"kubernetes.io/projected/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-kube-api-access-99twn\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.628092 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-catalog-content\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.730084 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-utilities\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.730237 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99twn\" (UniqueName: \"kubernetes.io/projected/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-kube-api-access-99twn\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.730302 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-catalog-content\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.730582 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-utilities\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.730781 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-catalog-content\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.751493 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99twn\" (UniqueName: \"kubernetes.io/projected/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-kube-api-access-99twn\") pod \"certified-operators-mrp9n\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:16 crc kubenswrapper[4809]: I1205 13:10:16.852096 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:17 crc kubenswrapper[4809]: I1205 13:10:17.397977 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrp9n"] Dec 05 13:10:17 crc kubenswrapper[4809]: I1205 13:10:17.988739 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrp9n" event={"ID":"a842d0e6-e09e-434e-9b8e-5a51a85fdda5","Type":"ContainerStarted","Data":"3fa7dfd103666f1243be2a5035a679d81c91d907f03ef7e8997432fbba300e11"} Dec 05 13:10:18 crc kubenswrapper[4809]: I1205 13:10:18.871985 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:10:18 crc kubenswrapper[4809]: E1205 13:10:18.873036 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:10:21 crc kubenswrapper[4809]: I1205 13:10:21.023940 4809 generic.go:334] "Generic (PLEG): container finished" podID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerID="6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829" exitCode=0 Dec 05 13:10:21 crc kubenswrapper[4809]: I1205 13:10:21.023989 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrp9n" event={"ID":"a842d0e6-e09e-434e-9b8e-5a51a85fdda5","Type":"ContainerDied","Data":"6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829"} Dec 05 13:10:23 crc kubenswrapper[4809]: I1205 13:10:23.046158 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrp9n" event={"ID":"a842d0e6-e09e-434e-9b8e-5a51a85fdda5","Type":"ContainerStarted","Data":"8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4"} Dec 05 13:10:24 crc kubenswrapper[4809]: I1205 13:10:24.057153 4809 generic.go:334] "Generic (PLEG): container finished" podID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerID="8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4" exitCode=0 Dec 05 13:10:24 crc kubenswrapper[4809]: I1205 13:10:24.057223 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrp9n" event={"ID":"a842d0e6-e09e-434e-9b8e-5a51a85fdda5","Type":"ContainerDied","Data":"8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4"} Dec 05 13:10:27 crc kubenswrapper[4809]: I1205 13:10:27.086391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrp9n" event={"ID":"a842d0e6-e09e-434e-9b8e-5a51a85fdda5","Type":"ContainerStarted","Data":"cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93"} Dec 05 13:10:27 crc kubenswrapper[4809]: I1205 13:10:27.114892 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mrp9n" podStartSLOduration=5.760898105 podStartE2EDuration="11.114804278s" podCreationTimestamp="2025-12-05 13:10:16 +0000 UTC" firstStartedPulling="2025-12-05 13:10:21.026250805 +0000 UTC m=+7316.417227373" lastFinishedPulling="2025-12-05 13:10:26.380156988 +0000 UTC m=+7321.771133546" observedRunningTime="2025-12-05 13:10:27.103912885 +0000 UTC m=+7322.494889443" watchObservedRunningTime="2025-12-05 13:10:27.114804278 +0000 UTC m=+7322.505780836" Dec 05 13:10:33 crc kubenswrapper[4809]: I1205 13:10:33.875086 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:10:33 crc kubenswrapper[4809]: E1205 13:10:33.876444 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:10:36 crc kubenswrapper[4809]: I1205 13:10:36.852789 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:36 crc kubenswrapper[4809]: I1205 13:10:36.853360 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:36 crc kubenswrapper[4809]: I1205 13:10:36.928584 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:37 crc kubenswrapper[4809]: I1205 13:10:37.221395 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:37 crc kubenswrapper[4809]: I1205 13:10:37.275489 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrp9n"] Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.201493 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mrp9n" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="registry-server" containerID="cri-o://cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93" gracePeriod=2 Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.721682 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.860352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99twn\" (UniqueName: \"kubernetes.io/projected/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-kube-api-access-99twn\") pod \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.860545 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-catalog-content\") pod \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.860588 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-utilities\") pod \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\" (UID: \"a842d0e6-e09e-434e-9b8e-5a51a85fdda5\") " Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.861482 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-utilities" (OuterVolumeSpecName: "utilities") pod "a842d0e6-e09e-434e-9b8e-5a51a85fdda5" (UID: "a842d0e6-e09e-434e-9b8e-5a51a85fdda5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.866724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-kube-api-access-99twn" (OuterVolumeSpecName: "kube-api-access-99twn") pod "a842d0e6-e09e-434e-9b8e-5a51a85fdda5" (UID: "a842d0e6-e09e-434e-9b8e-5a51a85fdda5"). InnerVolumeSpecName "kube-api-access-99twn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.908102 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a842d0e6-e09e-434e-9b8e-5a51a85fdda5" (UID: "a842d0e6-e09e-434e-9b8e-5a51a85fdda5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.963367 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.963579 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:10:39 crc kubenswrapper[4809]: I1205 13:10:39.963692 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99twn\" (UniqueName: \"kubernetes.io/projected/a842d0e6-e09e-434e-9b8e-5a51a85fdda5-kube-api-access-99twn\") on node \"crc\" DevicePath \"\"" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.213590 4809 generic.go:334] "Generic (PLEG): container finished" podID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerID="cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93" exitCode=0 Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.213657 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrp9n" event={"ID":"a842d0e6-e09e-434e-9b8e-5a51a85fdda5","Type":"ContainerDied","Data":"cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93"} Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.213690 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrp9n" event={"ID":"a842d0e6-e09e-434e-9b8e-5a51a85fdda5","Type":"ContainerDied","Data":"3fa7dfd103666f1243be2a5035a679d81c91d907f03ef7e8997432fbba300e11"} Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.213709 4809 scope.go:117] "RemoveContainer" containerID="cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.213803 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrp9n" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.238431 4809 scope.go:117] "RemoveContainer" containerID="8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.256759 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrp9n"] Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.265597 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mrp9n"] Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.267613 4809 scope.go:117] "RemoveContainer" containerID="6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.324398 4809 scope.go:117] "RemoveContainer" containerID="cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93" Dec 05 13:10:40 crc kubenswrapper[4809]: E1205 13:10:40.329429 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93\": container with ID starting with cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93 not found: ID does not exist" containerID="cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.329485 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93"} err="failed to get container status \"cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93\": rpc error: code = NotFound desc = could not find container \"cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93\": container with ID starting with cd95ae128d905100eab63d3a7bbcdb5bce563a485cd40ce4ce852432b8df7d93 not found: ID does not exist" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.329520 4809 scope.go:117] "RemoveContainer" containerID="8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4" Dec 05 13:10:40 crc kubenswrapper[4809]: E1205 13:10:40.330011 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4\": container with ID starting with 8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4 not found: ID does not exist" containerID="8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.330056 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4"} err="failed to get container status \"8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4\": rpc error: code = NotFound desc = could not find container \"8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4\": container with ID starting with 8c2f78455b6b39a3e5a9759dc7337593fc903dee67778828c4b55e06f41446b4 not found: ID does not exist" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.330083 4809 scope.go:117] "RemoveContainer" containerID="6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829" Dec 05 13:10:40 crc kubenswrapper[4809]: E1205 13:10:40.330582 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829\": container with ID starting with 6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829 not found: ID does not exist" containerID="6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.330601 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829"} err="failed to get container status \"6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829\": rpc error: code = NotFound desc = could not find container \"6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829\": container with ID starting with 6b868148e36bc4d91736f113557c74475e1167fc44d5a640efa3ed7e10709829 not found: ID does not exist" Dec 05 13:10:40 crc kubenswrapper[4809]: I1205 13:10:40.888157 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" path="/var/lib/kubelet/pods/a842d0e6-e09e-434e-9b8e-5a51a85fdda5/volumes" Dec 05 13:10:46 crc kubenswrapper[4809]: I1205 13:10:46.871933 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:10:46 crc kubenswrapper[4809]: E1205 13:10:46.872713 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:10:59 crc kubenswrapper[4809]: I1205 13:10:59.902816 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:10:59 crc kubenswrapper[4809]: E1205 13:10:59.903528 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:11:03 crc kubenswrapper[4809]: I1205 13:11:03.459906 4809 generic.go:334] "Generic (PLEG): container finished" podID="2ba9f48f-f049-46da-b82b-62785eff9fd3" containerID="b3ca428f1edf967449e79e1cd8c1defad4e0ea93b24d84655e90ef6af9c8fb38" exitCode=0 Dec 05 13:11:03 crc kubenswrapper[4809]: I1205 13:11:03.459989 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" event={"ID":"2ba9f48f-f049-46da-b82b-62785eff9fd3","Type":"ContainerDied","Data":"b3ca428f1edf967449e79e1cd8c1defad4e0ea93b24d84655e90ef6af9c8fb38"} Dec 05 13:11:04 crc kubenswrapper[4809]: I1205 13:11:04.968271 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.114920 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ssh-key\") pod \"2ba9f48f-f049-46da-b82b-62785eff9fd3\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.115198 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-inventory\") pod \"2ba9f48f-f049-46da-b82b-62785eff9fd3\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.115244 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ceph\") pod \"2ba9f48f-f049-46da-b82b-62785eff9fd3\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.115273 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk2jz\" (UniqueName: \"kubernetes.io/projected/2ba9f48f-f049-46da-b82b-62785eff9fd3-kube-api-access-kk2jz\") pod \"2ba9f48f-f049-46da-b82b-62785eff9fd3\" (UID: \"2ba9f48f-f049-46da-b82b-62785eff9fd3\") " Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.120943 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ceph" (OuterVolumeSpecName: "ceph") pod "2ba9f48f-f049-46da-b82b-62785eff9fd3" (UID: "2ba9f48f-f049-46da-b82b-62785eff9fd3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.121291 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ba9f48f-f049-46da-b82b-62785eff9fd3-kube-api-access-kk2jz" (OuterVolumeSpecName: "kube-api-access-kk2jz") pod "2ba9f48f-f049-46da-b82b-62785eff9fd3" (UID: "2ba9f48f-f049-46da-b82b-62785eff9fd3"). InnerVolumeSpecName "kube-api-access-kk2jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.143588 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ba9f48f-f049-46da-b82b-62785eff9fd3" (UID: "2ba9f48f-f049-46da-b82b-62785eff9fd3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.143830 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-inventory" (OuterVolumeSpecName: "inventory") pod "2ba9f48f-f049-46da-b82b-62785eff9fd3" (UID: "2ba9f48f-f049-46da-b82b-62785eff9fd3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.217577 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.217619 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.217633 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk2jz\" (UniqueName: \"kubernetes.io/projected/2ba9f48f-f049-46da-b82b-62785eff9fd3-kube-api-access-kk2jz\") on node \"crc\" DevicePath \"\"" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.217656 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ba9f48f-f049-46da-b82b-62785eff9fd3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.479653 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" event={"ID":"2ba9f48f-f049-46da-b82b-62785eff9fd3","Type":"ContainerDied","Data":"7a669fa03f221da58bea475cdda040bd8f5b3d5da4751f7f5da7b9297bfe85c6"} Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.479929 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a669fa03f221da58bea475cdda040bd8f5b3d5da4751f7f5da7b9297bfe85c6" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.479724 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-99s2k" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.566480 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-4r56d"] Dec 05 13:11:05 crc kubenswrapper[4809]: E1205 13:11:05.567088 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="extract-content" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.567115 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="extract-content" Dec 05 13:11:05 crc kubenswrapper[4809]: E1205 13:11:05.567155 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="registry-server" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.567164 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="registry-server" Dec 05 13:11:05 crc kubenswrapper[4809]: E1205 13:11:05.567190 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba9f48f-f049-46da-b82b-62785eff9fd3" containerName="download-cache-openstack-openstack-cell1" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.567203 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba9f48f-f049-46da-b82b-62785eff9fd3" containerName="download-cache-openstack-openstack-cell1" Dec 05 13:11:05 crc kubenswrapper[4809]: E1205 13:11:05.567217 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="extract-utilities" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.567226 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="extract-utilities" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.567513 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba9f48f-f049-46da-b82b-62785eff9fd3" containerName="download-cache-openstack-openstack-cell1" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.567546 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a842d0e6-e09e-434e-9b8e-5a51a85fdda5" containerName="registry-server" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.568774 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.571023 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.571289 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.571483 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.573100 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.578027 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-4r56d"] Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.728733 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87tp6\" (UniqueName: \"kubernetes.io/projected/0fbfa629-2f16-46c9-a67b-30b89444e688-kube-api-access-87tp6\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.728988 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ceph\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.729259 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-inventory\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.729364 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ssh-key\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.831533 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87tp6\" (UniqueName: \"kubernetes.io/projected/0fbfa629-2f16-46c9-a67b-30b89444e688-kube-api-access-87tp6\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.831666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ceph\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.831762 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-inventory\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.831820 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ssh-key\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.836920 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-inventory\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.836958 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ceph\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.841087 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ssh-key\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.850979 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87tp6\" (UniqueName: \"kubernetes.io/projected/0fbfa629-2f16-46c9-a67b-30b89444e688-kube-api-access-87tp6\") pod \"configure-network-openstack-openstack-cell1-4r56d\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:05 crc kubenswrapper[4809]: I1205 13:11:05.886827 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:11:06 crc kubenswrapper[4809]: I1205 13:11:06.434737 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-4r56d"] Dec 05 13:11:06 crc kubenswrapper[4809]: W1205 13:11:06.436108 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fbfa629_2f16_46c9_a67b_30b89444e688.slice/crio-661f9ec2f146cf03b346c9ec8e0c183e64d2c506ced51ed02f9f0beadc73bca3 WatchSource:0}: Error finding container 661f9ec2f146cf03b346c9ec8e0c183e64d2c506ced51ed02f9f0beadc73bca3: Status 404 returned error can't find the container with id 661f9ec2f146cf03b346c9ec8e0c183e64d2c506ced51ed02f9f0beadc73bca3 Dec 05 13:11:06 crc kubenswrapper[4809]: I1205 13:11:06.490350 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" event={"ID":"0fbfa629-2f16-46c9-a67b-30b89444e688","Type":"ContainerStarted","Data":"661f9ec2f146cf03b346c9ec8e0c183e64d2c506ced51ed02f9f0beadc73bca3"} Dec 05 13:11:07 crc kubenswrapper[4809]: I1205 13:11:07.501063 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" event={"ID":"0fbfa629-2f16-46c9-a67b-30b89444e688","Type":"ContainerStarted","Data":"3e9d99bf08547cd3825467ce9d0d6adb503b46e55585f7006b6eb718fd9c6653"} Dec 05 13:11:07 crc kubenswrapper[4809]: I1205 13:11:07.530118 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" podStartSLOduration=2.143411271 podStartE2EDuration="2.530081746s" podCreationTimestamp="2025-12-05 13:11:05 +0000 UTC" firstStartedPulling="2025-12-05 13:11:06.439052227 +0000 UTC m=+7361.830028785" lastFinishedPulling="2025-12-05 13:11:06.825722702 +0000 UTC m=+7362.216699260" observedRunningTime="2025-12-05 13:11:07.521197777 +0000 UTC m=+7362.912174355" watchObservedRunningTime="2025-12-05 13:11:07.530081746 +0000 UTC m=+7362.921058304" Dec 05 13:11:12 crc kubenswrapper[4809]: I1205 13:11:12.871992 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:11:12 crc kubenswrapper[4809]: E1205 13:11:12.873030 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:11:25 crc kubenswrapper[4809]: I1205 13:11:25.873429 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:11:25 crc kubenswrapper[4809]: E1205 13:11:25.875367 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:11:37 crc kubenswrapper[4809]: I1205 13:11:37.872990 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:11:37 crc kubenswrapper[4809]: E1205 13:11:37.873808 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:11:50 crc kubenswrapper[4809]: I1205 13:11:50.873180 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:11:51 crc kubenswrapper[4809]: I1205 13:11:51.990091 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"a1ad1f7dac7d9df8e8ac1f4d579d4e9cb79c07e17a81b299ada3af9bcee2479c"} Dec 05 13:12:28 crc kubenswrapper[4809]: I1205 13:12:28.387452 4809 generic.go:334] "Generic (PLEG): container finished" podID="0fbfa629-2f16-46c9-a67b-30b89444e688" containerID="3e9d99bf08547cd3825467ce9d0d6adb503b46e55585f7006b6eb718fd9c6653" exitCode=0 Dec 05 13:12:28 crc kubenswrapper[4809]: I1205 13:12:28.387623 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" event={"ID":"0fbfa629-2f16-46c9-a67b-30b89444e688","Type":"ContainerDied","Data":"3e9d99bf08547cd3825467ce9d0d6adb503b46e55585f7006b6eb718fd9c6653"} Dec 05 13:12:29 crc kubenswrapper[4809]: I1205 13:12:29.932938 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:12:29 crc kubenswrapper[4809]: I1205 13:12:29.992540 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87tp6\" (UniqueName: \"kubernetes.io/projected/0fbfa629-2f16-46c9-a67b-30b89444e688-kube-api-access-87tp6\") pod \"0fbfa629-2f16-46c9-a67b-30b89444e688\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " Dec 05 13:12:29 crc kubenswrapper[4809]: I1205 13:12:29.993220 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-inventory\") pod \"0fbfa629-2f16-46c9-a67b-30b89444e688\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " Dec 05 13:12:29 crc kubenswrapper[4809]: I1205 13:12:29.993413 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ceph\") pod \"0fbfa629-2f16-46c9-a67b-30b89444e688\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " Dec 05 13:12:29 crc kubenswrapper[4809]: I1205 13:12:29.993917 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ssh-key\") pod \"0fbfa629-2f16-46c9-a67b-30b89444e688\" (UID: \"0fbfa629-2f16-46c9-a67b-30b89444e688\") " Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.003051 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fbfa629-2f16-46c9-a67b-30b89444e688-kube-api-access-87tp6" (OuterVolumeSpecName: "kube-api-access-87tp6") pod "0fbfa629-2f16-46c9-a67b-30b89444e688" (UID: "0fbfa629-2f16-46c9-a67b-30b89444e688"). InnerVolumeSpecName "kube-api-access-87tp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.003190 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ceph" (OuterVolumeSpecName: "ceph") pod "0fbfa629-2f16-46c9-a67b-30b89444e688" (UID: "0fbfa629-2f16-46c9-a67b-30b89444e688"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.025003 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-inventory" (OuterVolumeSpecName: "inventory") pod "0fbfa629-2f16-46c9-a67b-30b89444e688" (UID: "0fbfa629-2f16-46c9-a67b-30b89444e688"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.030980 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0fbfa629-2f16-46c9-a67b-30b89444e688" (UID: "0fbfa629-2f16-46c9-a67b-30b89444e688"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.097588 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.097688 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.097707 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87tp6\" (UniqueName: \"kubernetes.io/projected/0fbfa629-2f16-46c9-a67b-30b89444e688-kube-api-access-87tp6\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.097716 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fbfa629-2f16-46c9-a67b-30b89444e688-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.407340 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" event={"ID":"0fbfa629-2f16-46c9-a67b-30b89444e688","Type":"ContainerDied","Data":"661f9ec2f146cf03b346c9ec8e0c183e64d2c506ced51ed02f9f0beadc73bca3"} Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.407729 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="661f9ec2f146cf03b346c9ec8e0c183e64d2c506ced51ed02f9f0beadc73bca3" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.407727 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-4r56d" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.510897 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-n4wl8"] Dec 05 13:12:30 crc kubenswrapper[4809]: E1205 13:12:30.511974 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbfa629-2f16-46c9-a67b-30b89444e688" containerName="configure-network-openstack-openstack-cell1" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.515694 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbfa629-2f16-46c9-a67b-30b89444e688" containerName="configure-network-openstack-openstack-cell1" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.516557 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbfa629-2f16-46c9-a67b-30b89444e688" containerName="configure-network-openstack-openstack-cell1" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.517964 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.522300 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.522889 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.523025 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.523153 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.526515 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-n4wl8"] Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.611560 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-inventory\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.611653 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7n62\" (UniqueName: \"kubernetes.io/projected/1e49a268-2565-4d37-aaa1-c256617a13a4-kube-api-access-z7n62\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.612112 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ssh-key\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.612384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ceph\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.714566 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-inventory\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.714649 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7n62\" (UniqueName: \"kubernetes.io/projected/1e49a268-2565-4d37-aaa1-c256617a13a4-kube-api-access-z7n62\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.714723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ssh-key\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.714769 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ceph\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.718711 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-inventory\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.718732 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ceph\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.726159 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ssh-key\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.748307 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7n62\" (UniqueName: \"kubernetes.io/projected/1e49a268-2565-4d37-aaa1-c256617a13a4-kube-api-access-z7n62\") pod \"validate-network-openstack-openstack-cell1-n4wl8\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:30 crc kubenswrapper[4809]: I1205 13:12:30.841404 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:31 crc kubenswrapper[4809]: I1205 13:12:31.484691 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-n4wl8"] Dec 05 13:12:32 crc kubenswrapper[4809]: I1205 13:12:32.428978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" event={"ID":"1e49a268-2565-4d37-aaa1-c256617a13a4","Type":"ContainerStarted","Data":"ad4cd96749f5386a150b0d2449f2cc72a038c8895be481db5b94d078b8a79ac1"} Dec 05 13:12:35 crc kubenswrapper[4809]: I1205 13:12:35.462530 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" event={"ID":"1e49a268-2565-4d37-aaa1-c256617a13a4","Type":"ContainerStarted","Data":"f25cddef31a829e25cdd7f0dffd35f330bff9f1c8a75246db88cd2fdfcee0b50"} Dec 05 13:12:40 crc kubenswrapper[4809]: I1205 13:12:40.514571 4809 generic.go:334] "Generic (PLEG): container finished" podID="1e49a268-2565-4d37-aaa1-c256617a13a4" containerID="f25cddef31a829e25cdd7f0dffd35f330bff9f1c8a75246db88cd2fdfcee0b50" exitCode=0 Dec 05 13:12:40 crc kubenswrapper[4809]: I1205 13:12:40.514689 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" event={"ID":"1e49a268-2565-4d37-aaa1-c256617a13a4","Type":"ContainerDied","Data":"f25cddef31a829e25cdd7f0dffd35f330bff9f1c8a75246db88cd2fdfcee0b50"} Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.770653 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n2sg2"] Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.773562 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.799367 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2sg2"] Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.884745 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-utilities\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.885125 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh2j5\" (UniqueName: \"kubernetes.io/projected/6d381b99-82fc-4955-822a-bc0ec6c28d57-kube-api-access-mh2j5\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.885177 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-catalog-content\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.987424 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-utilities\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.987556 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh2j5\" (UniqueName: \"kubernetes.io/projected/6d381b99-82fc-4955-822a-bc0ec6c28d57-kube-api-access-mh2j5\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.987653 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-utilities\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.987723 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-catalog-content\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:41 crc kubenswrapper[4809]: I1205 13:12:41.988099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-catalog-content\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.007683 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh2j5\" (UniqueName: \"kubernetes.io/projected/6d381b99-82fc-4955-822a-bc0ec6c28d57-kube-api-access-mh2j5\") pod \"redhat-marketplace-n2sg2\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.077056 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.108758 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.193160 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ssh-key\") pod \"1e49a268-2565-4d37-aaa1-c256617a13a4\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.193259 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7n62\" (UniqueName: \"kubernetes.io/projected/1e49a268-2565-4d37-aaa1-c256617a13a4-kube-api-access-z7n62\") pod \"1e49a268-2565-4d37-aaa1-c256617a13a4\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.193484 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-inventory\") pod \"1e49a268-2565-4d37-aaa1-c256617a13a4\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.193570 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ceph\") pod \"1e49a268-2565-4d37-aaa1-c256617a13a4\" (UID: \"1e49a268-2565-4d37-aaa1-c256617a13a4\") " Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.198366 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ceph" (OuterVolumeSpecName: "ceph") pod "1e49a268-2565-4d37-aaa1-c256617a13a4" (UID: "1e49a268-2565-4d37-aaa1-c256617a13a4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.199836 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e49a268-2565-4d37-aaa1-c256617a13a4-kube-api-access-z7n62" (OuterVolumeSpecName: "kube-api-access-z7n62") pod "1e49a268-2565-4d37-aaa1-c256617a13a4" (UID: "1e49a268-2565-4d37-aaa1-c256617a13a4"). InnerVolumeSpecName "kube-api-access-z7n62". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.238825 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-inventory" (OuterVolumeSpecName: "inventory") pod "1e49a268-2565-4d37-aaa1-c256617a13a4" (UID: "1e49a268-2565-4d37-aaa1-c256617a13a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.245354 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1e49a268-2565-4d37-aaa1-c256617a13a4" (UID: "1e49a268-2565-4d37-aaa1-c256617a13a4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.308523 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.308918 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7n62\" (UniqueName: \"kubernetes.io/projected/1e49a268-2565-4d37-aaa1-c256617a13a4-kube-api-access-z7n62\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.308931 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.308940 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e49a268-2565-4d37-aaa1-c256617a13a4-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.537945 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" event={"ID":"1e49a268-2565-4d37-aaa1-c256617a13a4","Type":"ContainerDied","Data":"ad4cd96749f5386a150b0d2449f2cc72a038c8895be481db5b94d078b8a79ac1"} Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.537998 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad4cd96749f5386a150b0d2449f2cc72a038c8895be481db5b94d078b8a79ac1" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.538074 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-n4wl8" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.633679 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-st5fn"] Dec 05 13:12:42 crc kubenswrapper[4809]: E1205 13:12:42.634151 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e49a268-2565-4d37-aaa1-c256617a13a4" containerName="validate-network-openstack-openstack-cell1" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.634171 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e49a268-2565-4d37-aaa1-c256617a13a4" containerName="validate-network-openstack-openstack-cell1" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.634393 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e49a268-2565-4d37-aaa1-c256617a13a4" containerName="validate-network-openstack-openstack-cell1" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.635358 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.638176 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.638203 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.638371 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.638522 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.652577 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-st5fn"] Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.701049 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2sg2"] Dec 05 13:12:42 crc kubenswrapper[4809]: W1205 13:12:42.702956 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d381b99_82fc_4955_822a_bc0ec6c28d57.slice/crio-7be496f97196a9993303e1ca0260659fa69b66cfe6d5a6648691ea47750eacf3 WatchSource:0}: Error finding container 7be496f97196a9993303e1ca0260659fa69b66cfe6d5a6648691ea47750eacf3: Status 404 returned error can't find the container with id 7be496f97196a9993303e1ca0260659fa69b66cfe6d5a6648691ea47750eacf3 Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.717888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ceph\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.718188 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-inventory\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.718323 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sp26\" (UniqueName: \"kubernetes.io/projected/8a9a80df-5202-4649-a667-f121649e4acc-kube-api-access-6sp26\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.718497 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ssh-key\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.821936 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ssh-key\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.822970 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ceph\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.823160 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-inventory\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.823260 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sp26\" (UniqueName: \"kubernetes.io/projected/8a9a80df-5202-4649-a667-f121649e4acc-kube-api-access-6sp26\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.828162 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-inventory\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.828182 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ceph\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.828748 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ssh-key\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.839431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sp26\" (UniqueName: \"kubernetes.io/projected/8a9a80df-5202-4649-a667-f121649e4acc-kube-api-access-6sp26\") pod \"install-os-openstack-openstack-cell1-st5fn\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:42 crc kubenswrapper[4809]: I1205 13:12:42.968062 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:12:43 crc kubenswrapper[4809]: W1205 13:12:43.515559 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a9a80df_5202_4649_a667_f121649e4acc.slice/crio-243be4db027a4b83c5392f5f8113c55dbfc337834bc84eb0468d615fc7c33101 WatchSource:0}: Error finding container 243be4db027a4b83c5392f5f8113c55dbfc337834bc84eb0468d615fc7c33101: Status 404 returned error can't find the container with id 243be4db027a4b83c5392f5f8113c55dbfc337834bc84eb0468d615fc7c33101 Dec 05 13:12:43 crc kubenswrapper[4809]: I1205 13:12:43.517393 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-st5fn"] Dec 05 13:12:43 crc kubenswrapper[4809]: I1205 13:12:43.550588 4809 generic.go:334] "Generic (PLEG): container finished" podID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerID="c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16" exitCode=0 Dec 05 13:12:43 crc kubenswrapper[4809]: I1205 13:12:43.550713 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2sg2" event={"ID":"6d381b99-82fc-4955-822a-bc0ec6c28d57","Type":"ContainerDied","Data":"c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16"} Dec 05 13:12:43 crc kubenswrapper[4809]: I1205 13:12:43.551044 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2sg2" event={"ID":"6d381b99-82fc-4955-822a-bc0ec6c28d57","Type":"ContainerStarted","Data":"7be496f97196a9993303e1ca0260659fa69b66cfe6d5a6648691ea47750eacf3"} Dec 05 13:12:43 crc kubenswrapper[4809]: I1205 13:12:43.553542 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-st5fn" event={"ID":"8a9a80df-5202-4649-a667-f121649e4acc","Type":"ContainerStarted","Data":"243be4db027a4b83c5392f5f8113c55dbfc337834bc84eb0468d615fc7c33101"} Dec 05 13:12:45 crc kubenswrapper[4809]: I1205 13:12:45.584283 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2sg2" event={"ID":"6d381b99-82fc-4955-822a-bc0ec6c28d57","Type":"ContainerStarted","Data":"042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597"} Dec 05 13:12:45 crc kubenswrapper[4809]: I1205 13:12:45.586923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-st5fn" event={"ID":"8a9a80df-5202-4649-a667-f121649e4acc","Type":"ContainerStarted","Data":"98b6c828acb05704cc9875cbbc2ed8efd7446233aa7fab594b09d260164ed871"} Dec 05 13:12:46 crc kubenswrapper[4809]: I1205 13:12:46.599015 4809 generic.go:334] "Generic (PLEG): container finished" podID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerID="042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597" exitCode=0 Dec 05 13:12:46 crc kubenswrapper[4809]: I1205 13:12:46.599089 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2sg2" event={"ID":"6d381b99-82fc-4955-822a-bc0ec6c28d57","Type":"ContainerDied","Data":"042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597"} Dec 05 13:12:46 crc kubenswrapper[4809]: I1205 13:12:46.650919 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-st5fn" podStartSLOduration=3.999235009 podStartE2EDuration="4.650897893s" podCreationTimestamp="2025-12-05 13:12:42 +0000 UTC" firstStartedPulling="2025-12-05 13:12:43.518343158 +0000 UTC m=+7458.909319716" lastFinishedPulling="2025-12-05 13:12:44.170006032 +0000 UTC m=+7459.560982600" observedRunningTime="2025-12-05 13:12:46.642444595 +0000 UTC m=+7462.033421163" watchObservedRunningTime="2025-12-05 13:12:46.650897893 +0000 UTC m=+7462.041874451" Dec 05 13:12:49 crc kubenswrapper[4809]: I1205 13:12:49.632023 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2sg2" event={"ID":"6d381b99-82fc-4955-822a-bc0ec6c28d57","Type":"ContainerStarted","Data":"c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361"} Dec 05 13:12:49 crc kubenswrapper[4809]: I1205 13:12:49.667923 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n2sg2" podStartSLOduration=4.292050908 podStartE2EDuration="8.667903394s" podCreationTimestamp="2025-12-05 13:12:41 +0000 UTC" firstStartedPulling="2025-12-05 13:12:43.552502988 +0000 UTC m=+7458.943479546" lastFinishedPulling="2025-12-05 13:12:47.928355474 +0000 UTC m=+7463.319332032" observedRunningTime="2025-12-05 13:12:49.655934622 +0000 UTC m=+7465.046911180" watchObservedRunningTime="2025-12-05 13:12:49.667903394 +0000 UTC m=+7465.058879952" Dec 05 13:12:52 crc kubenswrapper[4809]: I1205 13:12:52.109687 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:52 crc kubenswrapper[4809]: I1205 13:12:52.110193 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:52 crc kubenswrapper[4809]: I1205 13:12:52.164503 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.260220 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4dr6c"] Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.263546 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.309493 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dr6c"] Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.420598 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-utilities\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.421934 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-catalog-content\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.422152 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k57jh\" (UniqueName: \"kubernetes.io/projected/b5411b4a-cc04-47e4-92ff-a993254c4ab2-kube-api-access-k57jh\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.523980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k57jh\" (UniqueName: \"kubernetes.io/projected/b5411b4a-cc04-47e4-92ff-a993254c4ab2-kube-api-access-k57jh\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.524198 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-utilities\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.524247 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-catalog-content\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.524714 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-utilities\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.524773 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-catalog-content\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.557824 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k57jh\" (UniqueName: \"kubernetes.io/projected/b5411b4a-cc04-47e4-92ff-a993254c4ab2-kube-api-access-k57jh\") pod \"community-operators-4dr6c\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:57 crc kubenswrapper[4809]: I1205 13:12:57.593415 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:12:58 crc kubenswrapper[4809]: I1205 13:12:58.146852 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dr6c"] Dec 05 13:12:58 crc kubenswrapper[4809]: I1205 13:12:58.738770 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerStarted","Data":"4385ef76ba244cc386b74fbd4f2497af02ac52d85ee64fda2a49408f3d8337c6"} Dec 05 13:13:00 crc kubenswrapper[4809]: I1205 13:13:00.757801 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerStarted","Data":"aad5171492151c00f83fbf6af09c1d9adf6c2f738e3a844b45da091e2e7601a4"} Dec 05 13:13:01 crc kubenswrapper[4809]: I1205 13:13:01.770067 4809 generic.go:334] "Generic (PLEG): container finished" podID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerID="aad5171492151c00f83fbf6af09c1d9adf6c2f738e3a844b45da091e2e7601a4" exitCode=0 Dec 05 13:13:01 crc kubenswrapper[4809]: I1205 13:13:01.770127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerDied","Data":"aad5171492151c00f83fbf6af09c1d9adf6c2f738e3a844b45da091e2e7601a4"} Dec 05 13:13:02 crc kubenswrapper[4809]: I1205 13:13:02.189124 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:13:02 crc kubenswrapper[4809]: I1205 13:13:02.249721 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2sg2"] Dec 05 13:13:02 crc kubenswrapper[4809]: I1205 13:13:02.782868 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n2sg2" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="registry-server" containerID="cri-o://c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361" gracePeriod=2 Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.598395 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.685229 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-catalog-content\") pod \"6d381b99-82fc-4955-822a-bc0ec6c28d57\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.685501 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh2j5\" (UniqueName: \"kubernetes.io/projected/6d381b99-82fc-4955-822a-bc0ec6c28d57-kube-api-access-mh2j5\") pod \"6d381b99-82fc-4955-822a-bc0ec6c28d57\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.685716 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-utilities\") pod \"6d381b99-82fc-4955-822a-bc0ec6c28d57\" (UID: \"6d381b99-82fc-4955-822a-bc0ec6c28d57\") " Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.686720 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-utilities" (OuterVolumeSpecName: "utilities") pod "6d381b99-82fc-4955-822a-bc0ec6c28d57" (UID: "6d381b99-82fc-4955-822a-bc0ec6c28d57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.693938 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d381b99-82fc-4955-822a-bc0ec6c28d57-kube-api-access-mh2j5" (OuterVolumeSpecName: "kube-api-access-mh2j5") pod "6d381b99-82fc-4955-822a-bc0ec6c28d57" (UID: "6d381b99-82fc-4955-822a-bc0ec6c28d57"). InnerVolumeSpecName "kube-api-access-mh2j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.729027 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d381b99-82fc-4955-822a-bc0ec6c28d57" (UID: "6d381b99-82fc-4955-822a-bc0ec6c28d57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.788371 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.788397 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh2j5\" (UniqueName: \"kubernetes.io/projected/6d381b99-82fc-4955-822a-bc0ec6c28d57-kube-api-access-mh2j5\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.788407 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d381b99-82fc-4955-822a-bc0ec6c28d57-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.793609 4809 generic.go:334] "Generic (PLEG): container finished" podID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerID="c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361" exitCode=0 Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.793689 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2sg2" event={"ID":"6d381b99-82fc-4955-822a-bc0ec6c28d57","Type":"ContainerDied","Data":"c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361"} Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.793719 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2sg2" event={"ID":"6d381b99-82fc-4955-822a-bc0ec6c28d57","Type":"ContainerDied","Data":"7be496f97196a9993303e1ca0260659fa69b66cfe6d5a6648691ea47750eacf3"} Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.793739 4809 scope.go:117] "RemoveContainer" containerID="c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.793916 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2sg2" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.797316 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerStarted","Data":"aec888f7db72003762bb6f934776a0a129c074ddee2d48914ddee3ae14ace770"} Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.823667 4809 scope.go:117] "RemoveContainer" containerID="042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.860876 4809 scope.go:117] "RemoveContainer" containerID="c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.874180 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2sg2"] Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.887729 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2sg2"] Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.914607 4809 scope.go:117] "RemoveContainer" containerID="c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361" Dec 05 13:13:03 crc kubenswrapper[4809]: E1205 13:13:03.915240 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361\": container with ID starting with c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361 not found: ID does not exist" containerID="c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.915332 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361"} err="failed to get container status \"c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361\": rpc error: code = NotFound desc = could not find container \"c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361\": container with ID starting with c6243f5a368804807313246e8e2848dd893b86ce19bc4d9fe48617193fcff361 not found: ID does not exist" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.915367 4809 scope.go:117] "RemoveContainer" containerID="042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597" Dec 05 13:13:03 crc kubenswrapper[4809]: E1205 13:13:03.915683 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597\": container with ID starting with 042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597 not found: ID does not exist" containerID="042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.915721 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597"} err="failed to get container status \"042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597\": rpc error: code = NotFound desc = could not find container \"042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597\": container with ID starting with 042ed9ba2fb7fa9dfb994211bb1509fc7cdeee81acf76861726b2e5cf30ba597 not found: ID does not exist" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.915767 4809 scope.go:117] "RemoveContainer" containerID="c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16" Dec 05 13:13:03 crc kubenswrapper[4809]: E1205 13:13:03.916121 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16\": container with ID starting with c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16 not found: ID does not exist" containerID="c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16" Dec 05 13:13:03 crc kubenswrapper[4809]: I1205 13:13:03.916161 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16"} err="failed to get container status \"c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16\": rpc error: code = NotFound desc = could not find container \"c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16\": container with ID starting with c6bca03d1727f67a775585d8c2df5b08a766d62985c276febf0824be6e69cb16 not found: ID does not exist" Dec 05 13:13:04 crc kubenswrapper[4809]: I1205 13:13:04.810016 4809 generic.go:334] "Generic (PLEG): container finished" podID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerID="aec888f7db72003762bb6f934776a0a129c074ddee2d48914ddee3ae14ace770" exitCode=0 Dec 05 13:13:04 crc kubenswrapper[4809]: I1205 13:13:04.810372 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerDied","Data":"aec888f7db72003762bb6f934776a0a129c074ddee2d48914ddee3ae14ace770"} Dec 05 13:13:04 crc kubenswrapper[4809]: I1205 13:13:04.887079 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" path="/var/lib/kubelet/pods/6d381b99-82fc-4955-822a-bc0ec6c28d57/volumes" Dec 05 13:13:06 crc kubenswrapper[4809]: I1205 13:13:06.838780 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerStarted","Data":"09c8b91654e053f1fd68f4248b3cf29202743484534373e7869adb8dd28d907f"} Dec 05 13:13:06 crc kubenswrapper[4809]: I1205 13:13:06.861086 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4dr6c" podStartSLOduration=4.966121192 podStartE2EDuration="9.86106612s" podCreationTimestamp="2025-12-05 13:12:57 +0000 UTC" firstStartedPulling="2025-12-05 13:13:00.759558319 +0000 UTC m=+7476.150534877" lastFinishedPulling="2025-12-05 13:13:05.654503247 +0000 UTC m=+7481.045479805" observedRunningTime="2025-12-05 13:13:06.856915248 +0000 UTC m=+7482.247891816" watchObservedRunningTime="2025-12-05 13:13:06.86106612 +0000 UTC m=+7482.252042678" Dec 05 13:13:07 crc kubenswrapper[4809]: I1205 13:13:07.595112 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:13:07 crc kubenswrapper[4809]: I1205 13:13:07.595332 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:13:07 crc kubenswrapper[4809]: I1205 13:13:07.686341 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:13:17 crc kubenswrapper[4809]: I1205 13:13:17.645133 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:13:17 crc kubenswrapper[4809]: I1205 13:13:17.714988 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4dr6c"] Dec 05 13:13:17 crc kubenswrapper[4809]: I1205 13:13:17.950071 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4dr6c" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="registry-server" containerID="cri-o://09c8b91654e053f1fd68f4248b3cf29202743484534373e7869adb8dd28d907f" gracePeriod=2 Dec 05 13:13:18 crc kubenswrapper[4809]: I1205 13:13:18.968899 4809 generic.go:334] "Generic (PLEG): container finished" podID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerID="09c8b91654e053f1fd68f4248b3cf29202743484534373e7869adb8dd28d907f" exitCode=0 Dec 05 13:13:18 crc kubenswrapper[4809]: I1205 13:13:18.969365 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerDied","Data":"09c8b91654e053f1fd68f4248b3cf29202743484534373e7869adb8dd28d907f"} Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.218726 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.361722 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-catalog-content\") pod \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.361771 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-utilities\") pod \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.362118 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k57jh\" (UniqueName: \"kubernetes.io/projected/b5411b4a-cc04-47e4-92ff-a993254c4ab2-kube-api-access-k57jh\") pod \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\" (UID: \"b5411b4a-cc04-47e4-92ff-a993254c4ab2\") " Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.362578 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-utilities" (OuterVolumeSpecName: "utilities") pod "b5411b4a-cc04-47e4-92ff-a993254c4ab2" (UID: "b5411b4a-cc04-47e4-92ff-a993254c4ab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.363069 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.369331 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5411b4a-cc04-47e4-92ff-a993254c4ab2-kube-api-access-k57jh" (OuterVolumeSpecName: "kube-api-access-k57jh") pod "b5411b4a-cc04-47e4-92ff-a993254c4ab2" (UID: "b5411b4a-cc04-47e4-92ff-a993254c4ab2"). InnerVolumeSpecName "kube-api-access-k57jh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.428294 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5411b4a-cc04-47e4-92ff-a993254c4ab2" (UID: "b5411b4a-cc04-47e4-92ff-a993254c4ab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.465534 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k57jh\" (UniqueName: \"kubernetes.io/projected/b5411b4a-cc04-47e4-92ff-a993254c4ab2-kube-api-access-k57jh\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.465592 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5411b4a-cc04-47e4-92ff-a993254c4ab2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.987179 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dr6c" event={"ID":"b5411b4a-cc04-47e4-92ff-a993254c4ab2","Type":"ContainerDied","Data":"4385ef76ba244cc386b74fbd4f2497af02ac52d85ee64fda2a49408f3d8337c6"} Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.987244 4809 scope.go:117] "RemoveContainer" containerID="09c8b91654e053f1fd68f4248b3cf29202743484534373e7869adb8dd28d907f" Dec 05 13:13:19 crc kubenswrapper[4809]: I1205 13:13:19.987294 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dr6c" Dec 05 13:13:20 crc kubenswrapper[4809]: I1205 13:13:20.014987 4809 scope.go:117] "RemoveContainer" containerID="aec888f7db72003762bb6f934776a0a129c074ddee2d48914ddee3ae14ace770" Dec 05 13:13:20 crc kubenswrapper[4809]: I1205 13:13:20.045050 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4dr6c"] Dec 05 13:13:20 crc kubenswrapper[4809]: I1205 13:13:20.054313 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4dr6c"] Dec 05 13:13:20 crc kubenswrapper[4809]: I1205 13:13:20.066895 4809 scope.go:117] "RemoveContainer" containerID="aad5171492151c00f83fbf6af09c1d9adf6c2f738e3a844b45da091e2e7601a4" Dec 05 13:13:20 crc kubenswrapper[4809]: I1205 13:13:20.891486 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" path="/var/lib/kubelet/pods/b5411b4a-cc04-47e4-92ff-a993254c4ab2/volumes" Dec 05 13:13:30 crc kubenswrapper[4809]: I1205 13:13:30.108344 4809 generic.go:334] "Generic (PLEG): container finished" podID="8a9a80df-5202-4649-a667-f121649e4acc" containerID="98b6c828acb05704cc9875cbbc2ed8efd7446233aa7fab594b09d260164ed871" exitCode=0 Dec 05 13:13:30 crc kubenswrapper[4809]: I1205 13:13:30.108594 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-st5fn" event={"ID":"8a9a80df-5202-4649-a667-f121649e4acc","Type":"ContainerDied","Data":"98b6c828acb05704cc9875cbbc2ed8efd7446233aa7fab594b09d260164ed871"} Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.664387 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.755861 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ssh-key\") pod \"8a9a80df-5202-4649-a667-f121649e4acc\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.756303 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sp26\" (UniqueName: \"kubernetes.io/projected/8a9a80df-5202-4649-a667-f121649e4acc-kube-api-access-6sp26\") pod \"8a9a80df-5202-4649-a667-f121649e4acc\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.756457 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-inventory\") pod \"8a9a80df-5202-4649-a667-f121649e4acc\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.756547 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ceph\") pod \"8a9a80df-5202-4649-a667-f121649e4acc\" (UID: \"8a9a80df-5202-4649-a667-f121649e4acc\") " Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.766834 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ceph" (OuterVolumeSpecName: "ceph") pod "8a9a80df-5202-4649-a667-f121649e4acc" (UID: "8a9a80df-5202-4649-a667-f121649e4acc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.767154 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a9a80df-5202-4649-a667-f121649e4acc-kube-api-access-6sp26" (OuterVolumeSpecName: "kube-api-access-6sp26") pod "8a9a80df-5202-4649-a667-f121649e4acc" (UID: "8a9a80df-5202-4649-a667-f121649e4acc"). InnerVolumeSpecName "kube-api-access-6sp26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.798857 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a9a80df-5202-4649-a667-f121649e4acc" (UID: "8a9a80df-5202-4649-a667-f121649e4acc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.806014 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-inventory" (OuterVolumeSpecName: "inventory") pod "8a9a80df-5202-4649-a667-f121649e4acc" (UID: "8a9a80df-5202-4649-a667-f121649e4acc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.859377 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.859415 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sp26\" (UniqueName: \"kubernetes.io/projected/8a9a80df-5202-4649-a667-f121649e4acc-kube-api-access-6sp26\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.859426 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:31 crc kubenswrapper[4809]: I1205 13:13:31.859435 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8a9a80df-5202-4649-a667-f121649e4acc-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.135053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-st5fn" event={"ID":"8a9a80df-5202-4649-a667-f121649e4acc","Type":"ContainerDied","Data":"243be4db027a4b83c5392f5f8113c55dbfc337834bc84eb0468d615fc7c33101"} Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.135099 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="243be4db027a4b83c5392f5f8113c55dbfc337834bc84eb0468d615fc7c33101" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.135152 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-st5fn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.240544 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-bsvmn"] Dec 05 13:13:32 crc kubenswrapper[4809]: E1205 13:13:32.241097 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="extract-utilities" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241118 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="extract-utilities" Dec 05 13:13:32 crc kubenswrapper[4809]: E1205 13:13:32.241133 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="registry-server" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241141 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="registry-server" Dec 05 13:13:32 crc kubenswrapper[4809]: E1205 13:13:32.241160 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="extract-content" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241169 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="extract-content" Dec 05 13:13:32 crc kubenswrapper[4809]: E1205 13:13:32.241177 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="registry-server" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241182 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="registry-server" Dec 05 13:13:32 crc kubenswrapper[4809]: E1205 13:13:32.241194 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a9a80df-5202-4649-a667-f121649e4acc" containerName="install-os-openstack-openstack-cell1" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241200 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a9a80df-5202-4649-a667-f121649e4acc" containerName="install-os-openstack-openstack-cell1" Dec 05 13:13:32 crc kubenswrapper[4809]: E1205 13:13:32.241224 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="extract-content" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241230 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="extract-content" Dec 05 13:13:32 crc kubenswrapper[4809]: E1205 13:13:32.241248 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="extract-utilities" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241254 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="extract-utilities" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241450 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a9a80df-5202-4649-a667-f121649e4acc" containerName="install-os-openstack-openstack-cell1" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241462 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5411b4a-cc04-47e4-92ff-a993254c4ab2" containerName="registry-server" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.241485 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d381b99-82fc-4955-822a-bc0ec6c28d57" containerName="registry-server" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.242321 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.246358 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.247058 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.247612 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.249702 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.265003 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-bsvmn"] Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.370481 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkrqv\" (UniqueName: \"kubernetes.io/projected/cf82dcca-750d-458f-92ea-d4bc9b713f48-kube-api-access-tkrqv\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.370581 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ssh-key\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.370764 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ceph\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.370829 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-inventory\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.472856 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkrqv\" (UniqueName: \"kubernetes.io/projected/cf82dcca-750d-458f-92ea-d4bc9b713f48-kube-api-access-tkrqv\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.472940 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ssh-key\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.473038 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ceph\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.473075 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-inventory\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.478933 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ssh-key\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.490075 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-inventory\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.490169 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ceph\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.493059 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkrqv\" (UniqueName: \"kubernetes.io/projected/cf82dcca-750d-458f-92ea-d4bc9b713f48-kube-api-access-tkrqv\") pod \"configure-os-openstack-openstack-cell1-bsvmn\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:32 crc kubenswrapper[4809]: I1205 13:13:32.559532 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:13:33 crc kubenswrapper[4809]: I1205 13:13:33.105431 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-bsvmn"] Dec 05 13:13:33 crc kubenswrapper[4809]: I1205 13:13:33.145946 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" event={"ID":"cf82dcca-750d-458f-92ea-d4bc9b713f48","Type":"ContainerStarted","Data":"8dd32264b8fa02e3a9967d4a9f305b6fe0cb05d6ab49a7650dc245586bd52ab1"} Dec 05 13:13:34 crc kubenswrapper[4809]: I1205 13:13:34.157354 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" event={"ID":"cf82dcca-750d-458f-92ea-d4bc9b713f48","Type":"ContainerStarted","Data":"6fa639763b3e3022c3eec273e43dfadce895205197ea7f191c731fd169562112"} Dec 05 13:13:34 crc kubenswrapper[4809]: I1205 13:13:34.174778 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" podStartSLOduration=1.6516186990000001 podStartE2EDuration="2.174718399s" podCreationTimestamp="2025-12-05 13:13:32 +0000 UTC" firstStartedPulling="2025-12-05 13:13:33.113285547 +0000 UTC m=+7508.504262105" lastFinishedPulling="2025-12-05 13:13:33.636385247 +0000 UTC m=+7509.027361805" observedRunningTime="2025-12-05 13:13:34.172143869 +0000 UTC m=+7509.563120457" watchObservedRunningTime="2025-12-05 13:13:34.174718399 +0000 UTC m=+7509.565694957" Dec 05 13:14:14 crc kubenswrapper[4809]: I1205 13:14:14.046077 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:14:14 crc kubenswrapper[4809]: I1205 13:14:14.046742 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:14:20 crc kubenswrapper[4809]: I1205 13:14:20.654879 4809 generic.go:334] "Generic (PLEG): container finished" podID="cf82dcca-750d-458f-92ea-d4bc9b713f48" containerID="6fa639763b3e3022c3eec273e43dfadce895205197ea7f191c731fd169562112" exitCode=0 Dec 05 13:14:20 crc kubenswrapper[4809]: I1205 13:14:20.654970 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" event={"ID":"cf82dcca-750d-458f-92ea-d4bc9b713f48","Type":"ContainerDied","Data":"6fa639763b3e3022c3eec273e43dfadce895205197ea7f191c731fd169562112"} Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.497354 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.626252 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ssh-key\") pod \"cf82dcca-750d-458f-92ea-d4bc9b713f48\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.626328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-inventory\") pod \"cf82dcca-750d-458f-92ea-d4bc9b713f48\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.626830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ceph\") pod \"cf82dcca-750d-458f-92ea-d4bc9b713f48\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.626908 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkrqv\" (UniqueName: \"kubernetes.io/projected/cf82dcca-750d-458f-92ea-d4bc9b713f48-kube-api-access-tkrqv\") pod \"cf82dcca-750d-458f-92ea-d4bc9b713f48\" (UID: \"cf82dcca-750d-458f-92ea-d4bc9b713f48\") " Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.652140 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf82dcca-750d-458f-92ea-d4bc9b713f48-kube-api-access-tkrqv" (OuterVolumeSpecName: "kube-api-access-tkrqv") pod "cf82dcca-750d-458f-92ea-d4bc9b713f48" (UID: "cf82dcca-750d-458f-92ea-d4bc9b713f48"). InnerVolumeSpecName "kube-api-access-tkrqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.655032 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkrqv\" (UniqueName: \"kubernetes.io/projected/cf82dcca-750d-458f-92ea-d4bc9b713f48-kube-api-access-tkrqv\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.678835 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ceph" (OuterVolumeSpecName: "ceph") pod "cf82dcca-750d-458f-92ea-d4bc9b713f48" (UID: "cf82dcca-750d-458f-92ea-d4bc9b713f48"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.710557 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" event={"ID":"cf82dcca-750d-458f-92ea-d4bc9b713f48","Type":"ContainerDied","Data":"8dd32264b8fa02e3a9967d4a9f305b6fe0cb05d6ab49a7650dc245586bd52ab1"} Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.710599 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dd32264b8fa02e3a9967d4a9f305b6fe0cb05d6ab49a7650dc245586bd52ab1" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.710692 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-bsvmn" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.764759 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-inventory" (OuterVolumeSpecName: "inventory") pod "cf82dcca-750d-458f-92ea-d4bc9b713f48" (UID: "cf82dcca-750d-458f-92ea-d4bc9b713f48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.770224 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.770269 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:22 crc kubenswrapper[4809]: I1205 13:14:22.779829 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cf82dcca-750d-458f-92ea-d4bc9b713f48" (UID: "cf82dcca-750d-458f-92ea-d4bc9b713f48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:22.874162 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf82dcca-750d-458f-92ea-d4bc9b713f48-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:22.956142 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-nd5b4"] Dec 05 13:14:23 crc kubenswrapper[4809]: E1205 13:14:22.957175 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf82dcca-750d-458f-92ea-d4bc9b713f48" containerName="configure-os-openstack-openstack-cell1" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:22.957196 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf82dcca-750d-458f-92ea-d4bc9b713f48" containerName="configure-os-openstack-openstack-cell1" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:22.957515 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf82dcca-750d-458f-92ea-d4bc9b713f48" containerName="configure-os-openstack-openstack-cell1" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:22.958814 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:22.967761 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-nd5b4"] Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.081002 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.081134 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ceph\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.081232 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vcn8\" (UniqueName: \"kubernetes.io/projected/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-kube-api-access-7vcn8\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.081269 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-inventory-0\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.184237 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vcn8\" (UniqueName: \"kubernetes.io/projected/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-kube-api-access-7vcn8\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.184337 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-inventory-0\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.184517 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.184652 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ceph\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.190425 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-inventory-0\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.190538 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ceph\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.191879 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.216259 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vcn8\" (UniqueName: \"kubernetes.io/projected/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-kube-api-access-7vcn8\") pod \"ssh-known-hosts-openstack-nd5b4\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:23 crc kubenswrapper[4809]: I1205 13:14:23.306274 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:24 crc kubenswrapper[4809]: I1205 13:14:24.064501 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-nd5b4"] Dec 05 13:14:24 crc kubenswrapper[4809]: I1205 13:14:24.730893 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-nd5b4" event={"ID":"bf168ba6-afe9-4d8e-b73f-1454c4ee6522","Type":"ContainerStarted","Data":"fb329dbf4292a80912fa27dad0707a47c05c50b1c0e962ac3b86b977cbd909c4"} Dec 05 13:14:25 crc kubenswrapper[4809]: I1205 13:14:25.740566 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-nd5b4" event={"ID":"bf168ba6-afe9-4d8e-b73f-1454c4ee6522","Type":"ContainerStarted","Data":"63b49b597ede857d7ffc1249451ac22fece39d93b8c5728714e198ea47ddd8fb"} Dec 05 13:14:25 crc kubenswrapper[4809]: I1205 13:14:25.772687 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-nd5b4" podStartSLOduration=2.8198001059999998 podStartE2EDuration="3.772639733s" podCreationTimestamp="2025-12-05 13:14:22 +0000 UTC" firstStartedPulling="2025-12-05 13:14:24.079261427 +0000 UTC m=+7559.470237985" lastFinishedPulling="2025-12-05 13:14:25.032101044 +0000 UTC m=+7560.423077612" observedRunningTime="2025-12-05 13:14:25.7665815 +0000 UTC m=+7561.157558068" watchObservedRunningTime="2025-12-05 13:14:25.772639733 +0000 UTC m=+7561.163616291" Dec 05 13:14:34 crc kubenswrapper[4809]: I1205 13:14:34.836662 4809 generic.go:334] "Generic (PLEG): container finished" podID="bf168ba6-afe9-4d8e-b73f-1454c4ee6522" containerID="63b49b597ede857d7ffc1249451ac22fece39d93b8c5728714e198ea47ddd8fb" exitCode=0 Dec 05 13:14:34 crc kubenswrapper[4809]: I1205 13:14:34.836738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-nd5b4" event={"ID":"bf168ba6-afe9-4d8e-b73f-1454c4ee6522","Type":"ContainerDied","Data":"63b49b597ede857d7ffc1249451ac22fece39d93b8c5728714e198ea47ddd8fb"} Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.513791 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.599823 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vcn8\" (UniqueName: \"kubernetes.io/projected/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-kube-api-access-7vcn8\") pod \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.599947 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-inventory-0\") pod \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.600008 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ssh-key-openstack-cell1\") pod \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.600247 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ceph\") pod \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\" (UID: \"bf168ba6-afe9-4d8e-b73f-1454c4ee6522\") " Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.606420 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ceph" (OuterVolumeSpecName: "ceph") pod "bf168ba6-afe9-4d8e-b73f-1454c4ee6522" (UID: "bf168ba6-afe9-4d8e-b73f-1454c4ee6522"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.607492 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-kube-api-access-7vcn8" (OuterVolumeSpecName: "kube-api-access-7vcn8") pod "bf168ba6-afe9-4d8e-b73f-1454c4ee6522" (UID: "bf168ba6-afe9-4d8e-b73f-1454c4ee6522"). InnerVolumeSpecName "kube-api-access-7vcn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.637979 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "bf168ba6-afe9-4d8e-b73f-1454c4ee6522" (UID: "bf168ba6-afe9-4d8e-b73f-1454c4ee6522"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.648317 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "bf168ba6-afe9-4d8e-b73f-1454c4ee6522" (UID: "bf168ba6-afe9-4d8e-b73f-1454c4ee6522"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.702883 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.703229 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vcn8\" (UniqueName: \"kubernetes.io/projected/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-kube-api-access-7vcn8\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.703241 4809 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.703251 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bf168ba6-afe9-4d8e-b73f-1454c4ee6522-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.857590 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-nd5b4" event={"ID":"bf168ba6-afe9-4d8e-b73f-1454c4ee6522","Type":"ContainerDied","Data":"fb329dbf4292a80912fa27dad0707a47c05c50b1c0e962ac3b86b977cbd909c4"} Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.857691 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb329dbf4292a80912fa27dad0707a47c05c50b1c0e962ac3b86b977cbd909c4" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.857729 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-nd5b4" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.933045 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-fwxrn"] Dec 05 13:14:36 crc kubenswrapper[4809]: E1205 13:14:36.933794 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf168ba6-afe9-4d8e-b73f-1454c4ee6522" containerName="ssh-known-hosts-openstack" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.933819 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf168ba6-afe9-4d8e-b73f-1454c4ee6522" containerName="ssh-known-hosts-openstack" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.934106 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf168ba6-afe9-4d8e-b73f-1454c4ee6522" containerName="ssh-known-hosts-openstack" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.935291 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.938334 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.938662 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.939195 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.939277 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:14:36 crc kubenswrapper[4809]: I1205 13:14:36.952265 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-fwxrn"] Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.009751 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-inventory\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.009944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnnsh\" (UniqueName: \"kubernetes.io/projected/43079fec-0ad2-472e-af20-6b102ed4284f-kube-api-access-dnnsh\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.010051 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ceph\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.010082 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ssh-key\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.112040 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnnsh\" (UniqueName: \"kubernetes.io/projected/43079fec-0ad2-472e-af20-6b102ed4284f-kube-api-access-dnnsh\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.112200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ceph\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.112242 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ssh-key\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.112279 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-inventory\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.129423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-inventory\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.129571 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ceph\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.129600 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ssh-key\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.133991 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnnsh\" (UniqueName: \"kubernetes.io/projected/43079fec-0ad2-472e-af20-6b102ed4284f-kube-api-access-dnnsh\") pod \"run-os-openstack-openstack-cell1-fwxrn\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.254789 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.838297 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-fwxrn"] Dec 05 13:14:37 crc kubenswrapper[4809]: W1205 13:14:37.843787 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43079fec_0ad2_472e_af20_6b102ed4284f.slice/crio-2fb817939c930dfa26a995500b7d96f877e1ddbb255889a63a39f8acf29a830f WatchSource:0}: Error finding container 2fb817939c930dfa26a995500b7d96f877e1ddbb255889a63a39f8acf29a830f: Status 404 returned error can't find the container with id 2fb817939c930dfa26a995500b7d96f877e1ddbb255889a63a39f8acf29a830f Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.847309 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:14:37 crc kubenswrapper[4809]: I1205 13:14:37.868923 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" event={"ID":"43079fec-0ad2-472e-af20-6b102ed4284f","Type":"ContainerStarted","Data":"2fb817939c930dfa26a995500b7d96f877e1ddbb255889a63a39f8acf29a830f"} Dec 05 13:14:38 crc kubenswrapper[4809]: I1205 13:14:38.885436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" event={"ID":"43079fec-0ad2-472e-af20-6b102ed4284f","Type":"ContainerStarted","Data":"ac12f66a79ec8d4b8890ff3e7727ebbf458e305a0484d461082183a5780fac51"} Dec 05 13:14:38 crc kubenswrapper[4809]: I1205 13:14:38.903050 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" podStartSLOduration=2.415847013 podStartE2EDuration="2.903031587s" podCreationTimestamp="2025-12-05 13:14:36 +0000 UTC" firstStartedPulling="2025-12-05 13:14:37.847082031 +0000 UTC m=+7573.238058589" lastFinishedPulling="2025-12-05 13:14:38.334266605 +0000 UTC m=+7573.725243163" observedRunningTime="2025-12-05 13:14:38.900469428 +0000 UTC m=+7574.291445986" watchObservedRunningTime="2025-12-05 13:14:38.903031587 +0000 UTC m=+7574.294008145" Dec 05 13:14:44 crc kubenswrapper[4809]: I1205 13:14:44.046464 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:14:44 crc kubenswrapper[4809]: I1205 13:14:44.046923 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:14:47 crc kubenswrapper[4809]: I1205 13:14:47.967099 4809 generic.go:334] "Generic (PLEG): container finished" podID="43079fec-0ad2-472e-af20-6b102ed4284f" containerID="ac12f66a79ec8d4b8890ff3e7727ebbf458e305a0484d461082183a5780fac51" exitCode=0 Dec 05 13:14:47 crc kubenswrapper[4809]: I1205 13:14:47.967355 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" event={"ID":"43079fec-0ad2-472e-af20-6b102ed4284f","Type":"ContainerDied","Data":"ac12f66a79ec8d4b8890ff3e7727ebbf458e305a0484d461082183a5780fac51"} Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.600591 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.690181 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnnsh\" (UniqueName: \"kubernetes.io/projected/43079fec-0ad2-472e-af20-6b102ed4284f-kube-api-access-dnnsh\") pod \"43079fec-0ad2-472e-af20-6b102ed4284f\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.690361 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ceph\") pod \"43079fec-0ad2-472e-af20-6b102ed4284f\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.690401 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ssh-key\") pod \"43079fec-0ad2-472e-af20-6b102ed4284f\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.690509 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-inventory\") pod \"43079fec-0ad2-472e-af20-6b102ed4284f\" (UID: \"43079fec-0ad2-472e-af20-6b102ed4284f\") " Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.711775 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ceph" (OuterVolumeSpecName: "ceph") pod "43079fec-0ad2-472e-af20-6b102ed4284f" (UID: "43079fec-0ad2-472e-af20-6b102ed4284f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.713094 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43079fec-0ad2-472e-af20-6b102ed4284f-kube-api-access-dnnsh" (OuterVolumeSpecName: "kube-api-access-dnnsh") pod "43079fec-0ad2-472e-af20-6b102ed4284f" (UID: "43079fec-0ad2-472e-af20-6b102ed4284f"). InnerVolumeSpecName "kube-api-access-dnnsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.739759 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-inventory" (OuterVolumeSpecName: "inventory") pod "43079fec-0ad2-472e-af20-6b102ed4284f" (UID: "43079fec-0ad2-472e-af20-6b102ed4284f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.771926 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43079fec-0ad2-472e-af20-6b102ed4284f" (UID: "43079fec-0ad2-472e-af20-6b102ed4284f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.794297 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.794343 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnnsh\" (UniqueName: \"kubernetes.io/projected/43079fec-0ad2-472e-af20-6b102ed4284f-kube-api-access-dnnsh\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.794357 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.794367 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43079fec-0ad2-472e-af20-6b102ed4284f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.986241 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" event={"ID":"43079fec-0ad2-472e-af20-6b102ed4284f","Type":"ContainerDied","Data":"2fb817939c930dfa26a995500b7d96f877e1ddbb255889a63a39f8acf29a830f"} Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.986283 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-fwxrn" Dec 05 13:14:49 crc kubenswrapper[4809]: I1205 13:14:49.986286 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb817939c930dfa26a995500b7d96f877e1ddbb255889a63a39f8acf29a830f" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.070886 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-gqxn9"] Dec 05 13:14:50 crc kubenswrapper[4809]: E1205 13:14:50.071394 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43079fec-0ad2-472e-af20-6b102ed4284f" containerName="run-os-openstack-openstack-cell1" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.071415 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="43079fec-0ad2-472e-af20-6b102ed4284f" containerName="run-os-openstack-openstack-cell1" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.071730 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="43079fec-0ad2-472e-af20-6b102ed4284f" containerName="run-os-openstack-openstack-cell1" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.072730 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.075303 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.075338 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.075550 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.076421 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.080566 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-gqxn9"] Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.201391 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.201502 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mx52\" (UniqueName: \"kubernetes.io/projected/386049c0-b517-465a-a3b8-283127b5de78-kube-api-access-9mx52\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.201601 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-inventory\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.201738 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ceph\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.303582 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.303695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mx52\" (UniqueName: \"kubernetes.io/projected/386049c0-b517-465a-a3b8-283127b5de78-kube-api-access-9mx52\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.303819 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-inventory\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.303890 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ceph\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.310433 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ceph\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.311570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.326580 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-inventory\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.328611 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mx52\" (UniqueName: \"kubernetes.io/projected/386049c0-b517-465a-a3b8-283127b5de78-kube-api-access-9mx52\") pod \"reboot-os-openstack-openstack-cell1-gqxn9\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.387872 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:14:50 crc kubenswrapper[4809]: I1205 13:14:50.980012 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-gqxn9"] Dec 05 13:14:51 crc kubenswrapper[4809]: I1205 13:14:51.000798 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" event={"ID":"386049c0-b517-465a-a3b8-283127b5de78","Type":"ContainerStarted","Data":"b300c11390199aabe9ac60e1260aff8871f400594a952cb4c77e9cd09d1cc63a"} Dec 05 13:14:52 crc kubenswrapper[4809]: I1205 13:14:52.014154 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" event={"ID":"386049c0-b517-465a-a3b8-283127b5de78","Type":"ContainerStarted","Data":"3c6372ddda6a7ac38ce971a639d60b6e2468735dcacd2668c434f44bb5ac9a8a"} Dec 05 13:14:52 crc kubenswrapper[4809]: I1205 13:14:52.052068 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" podStartSLOduration=1.45580371 podStartE2EDuration="2.052039412s" podCreationTimestamp="2025-12-05 13:14:50 +0000 UTC" firstStartedPulling="2025-12-05 13:14:50.982807049 +0000 UTC m=+7586.373783617" lastFinishedPulling="2025-12-05 13:14:51.579042721 +0000 UTC m=+7586.970019319" observedRunningTime="2025-12-05 13:14:52.042962868 +0000 UTC m=+7587.433939426" watchObservedRunningTime="2025-12-05 13:14:52.052039412 +0000 UTC m=+7587.443015970" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.160533 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn"] Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.164196 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.176795 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.177248 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.181461 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn"] Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.226674 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp9sc\" (UniqueName: \"kubernetes.io/projected/59fcdcce-9ede-4874-912d-c22ed395fa2d-kube-api-access-pp9sc\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.226809 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fcdcce-9ede-4874-912d-c22ed395fa2d-config-volume\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.226885 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fcdcce-9ede-4874-912d-c22ed395fa2d-secret-volume\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.329416 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp9sc\" (UniqueName: \"kubernetes.io/projected/59fcdcce-9ede-4874-912d-c22ed395fa2d-kube-api-access-pp9sc\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.329474 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fcdcce-9ede-4874-912d-c22ed395fa2d-config-volume\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.329497 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fcdcce-9ede-4874-912d-c22ed395fa2d-secret-volume\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.330751 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fcdcce-9ede-4874-912d-c22ed395fa2d-config-volume\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.336070 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fcdcce-9ede-4874-912d-c22ed395fa2d-secret-volume\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.350249 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp9sc\" (UniqueName: \"kubernetes.io/projected/59fcdcce-9ede-4874-912d-c22ed395fa2d-kube-api-access-pp9sc\") pod \"collect-profiles-29415675-qzxkn\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.496007 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:00 crc kubenswrapper[4809]: I1205 13:15:00.998916 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn"] Dec 05 13:15:01 crc kubenswrapper[4809]: I1205 13:15:01.098363 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" event={"ID":"59fcdcce-9ede-4874-912d-c22ed395fa2d","Type":"ContainerStarted","Data":"30fc16e53bc10102a4c09d27a1ce197ea7a2b24b40efd1bc92dce10373d0adf7"} Dec 05 13:15:02 crc kubenswrapper[4809]: I1205 13:15:02.108277 4809 generic.go:334] "Generic (PLEG): container finished" podID="59fcdcce-9ede-4874-912d-c22ed395fa2d" containerID="6d55b7fd746429610367ce5d8a919c4e5fb87868396e626cf584a6af4996839d" exitCode=0 Dec 05 13:15:02 crc kubenswrapper[4809]: I1205 13:15:02.108332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" event={"ID":"59fcdcce-9ede-4874-912d-c22ed395fa2d","Type":"ContainerDied","Data":"6d55b7fd746429610367ce5d8a919c4e5fb87868396e626cf584a6af4996839d"} Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.595049 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.606699 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fcdcce-9ede-4874-912d-c22ed395fa2d-config-volume\") pod \"59fcdcce-9ede-4874-912d-c22ed395fa2d\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.606776 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp9sc\" (UniqueName: \"kubernetes.io/projected/59fcdcce-9ede-4874-912d-c22ed395fa2d-kube-api-access-pp9sc\") pod \"59fcdcce-9ede-4874-912d-c22ed395fa2d\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.607265 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fcdcce-9ede-4874-912d-c22ed395fa2d-secret-volume\") pod \"59fcdcce-9ede-4874-912d-c22ed395fa2d\" (UID: \"59fcdcce-9ede-4874-912d-c22ed395fa2d\") " Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.607660 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fcdcce-9ede-4874-912d-c22ed395fa2d-config-volume" (OuterVolumeSpecName: "config-volume") pod "59fcdcce-9ede-4874-912d-c22ed395fa2d" (UID: "59fcdcce-9ede-4874-912d-c22ed395fa2d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.608517 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fcdcce-9ede-4874-912d-c22ed395fa2d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.614049 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59fcdcce-9ede-4874-912d-c22ed395fa2d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59fcdcce-9ede-4874-912d-c22ed395fa2d" (UID: "59fcdcce-9ede-4874-912d-c22ed395fa2d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.642157 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59fcdcce-9ede-4874-912d-c22ed395fa2d-kube-api-access-pp9sc" (OuterVolumeSpecName: "kube-api-access-pp9sc") pod "59fcdcce-9ede-4874-912d-c22ed395fa2d" (UID: "59fcdcce-9ede-4874-912d-c22ed395fa2d"). InnerVolumeSpecName "kube-api-access-pp9sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.711055 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fcdcce-9ede-4874-912d-c22ed395fa2d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:03 crc kubenswrapper[4809]: I1205 13:15:03.711103 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp9sc\" (UniqueName: \"kubernetes.io/projected/59fcdcce-9ede-4874-912d-c22ed395fa2d-kube-api-access-pp9sc\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:04 crc kubenswrapper[4809]: I1205 13:15:04.132048 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" event={"ID":"59fcdcce-9ede-4874-912d-c22ed395fa2d","Type":"ContainerDied","Data":"30fc16e53bc10102a4c09d27a1ce197ea7a2b24b40efd1bc92dce10373d0adf7"} Dec 05 13:15:04 crc kubenswrapper[4809]: I1205 13:15:04.132390 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30fc16e53bc10102a4c09d27a1ce197ea7a2b24b40efd1bc92dce10373d0adf7" Dec 05 13:15:04 crc kubenswrapper[4809]: I1205 13:15:04.132116 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn" Dec 05 13:15:04 crc kubenswrapper[4809]: I1205 13:15:04.675784 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5"] Dec 05 13:15:04 crc kubenswrapper[4809]: I1205 13:15:04.691862 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415630-bh8n5"] Dec 05 13:15:04 crc kubenswrapper[4809]: I1205 13:15:04.891452 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130471fd-53bd-4db9-9ba0-76e55f3778a2" path="/var/lib/kubelet/pods/130471fd-53bd-4db9-9ba0-76e55f3778a2/volumes" Dec 05 13:15:08 crc kubenswrapper[4809]: I1205 13:15:08.178736 4809 generic.go:334] "Generic (PLEG): container finished" podID="386049c0-b517-465a-a3b8-283127b5de78" containerID="3c6372ddda6a7ac38ce971a639d60b6e2468735dcacd2668c434f44bb5ac9a8a" exitCode=0 Dec 05 13:15:08 crc kubenswrapper[4809]: I1205 13:15:08.178854 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" event={"ID":"386049c0-b517-465a-a3b8-283127b5de78","Type":"ContainerDied","Data":"3c6372ddda6a7ac38ce971a639d60b6e2468735dcacd2668c434f44bb5ac9a8a"} Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.740260 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.759445 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mx52\" (UniqueName: \"kubernetes.io/projected/386049c0-b517-465a-a3b8-283127b5de78-kube-api-access-9mx52\") pod \"386049c0-b517-465a-a3b8-283127b5de78\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.759981 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-inventory\") pod \"386049c0-b517-465a-a3b8-283127b5de78\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.760070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ceph\") pod \"386049c0-b517-465a-a3b8-283127b5de78\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.760213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ssh-key\") pod \"386049c0-b517-465a-a3b8-283127b5de78\" (UID: \"386049c0-b517-465a-a3b8-283127b5de78\") " Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.765365 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ceph" (OuterVolumeSpecName: "ceph") pod "386049c0-b517-465a-a3b8-283127b5de78" (UID: "386049c0-b517-465a-a3b8-283127b5de78"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.765976 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/386049c0-b517-465a-a3b8-283127b5de78-kube-api-access-9mx52" (OuterVolumeSpecName: "kube-api-access-9mx52") pod "386049c0-b517-465a-a3b8-283127b5de78" (UID: "386049c0-b517-465a-a3b8-283127b5de78"). InnerVolumeSpecName "kube-api-access-9mx52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.799088 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-inventory" (OuterVolumeSpecName: "inventory") pod "386049c0-b517-465a-a3b8-283127b5de78" (UID: "386049c0-b517-465a-a3b8-283127b5de78"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.802422 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "386049c0-b517-465a-a3b8-283127b5de78" (UID: "386049c0-b517-465a-a3b8-283127b5de78"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.864570 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.864805 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.864821 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/386049c0-b517-465a-a3b8-283127b5de78-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:09 crc kubenswrapper[4809]: I1205 13:15:09.864839 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mx52\" (UniqueName: \"kubernetes.io/projected/386049c0-b517-465a-a3b8-283127b5de78-kube-api-access-9mx52\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.202742 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" event={"ID":"386049c0-b517-465a-a3b8-283127b5de78","Type":"ContainerDied","Data":"b300c11390199aabe9ac60e1260aff8871f400594a952cb4c77e9cd09d1cc63a"} Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.202791 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b300c11390199aabe9ac60e1260aff8871f400594a952cb4c77e9cd09d1cc63a" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.202802 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-gqxn9" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.293784 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-l9x7s"] Dec 05 13:15:10 crc kubenswrapper[4809]: E1205 13:15:10.294347 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59fcdcce-9ede-4874-912d-c22ed395fa2d" containerName="collect-profiles" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.294372 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="59fcdcce-9ede-4874-912d-c22ed395fa2d" containerName="collect-profiles" Dec 05 13:15:10 crc kubenswrapper[4809]: E1205 13:15:10.294418 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386049c0-b517-465a-a3b8-283127b5de78" containerName="reboot-os-openstack-openstack-cell1" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.294445 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="386049c0-b517-465a-a3b8-283127b5de78" containerName="reboot-os-openstack-openstack-cell1" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.294751 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="59fcdcce-9ede-4874-912d-c22ed395fa2d" containerName="collect-profiles" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.294774 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="386049c0-b517-465a-a3b8-283127b5de78" containerName="reboot-os-openstack-openstack-cell1" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.295800 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.299561 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.299598 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.300722 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.301239 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.307596 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-l9x7s"] Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.375085 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.375337 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmtnf\" (UniqueName: \"kubernetes.io/projected/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-kube-api-access-cmtnf\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.375503 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.375664 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.375778 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.375851 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-inventory\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.375944 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.376055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ssh-key\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.376172 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.376280 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.376374 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.376464 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ceph\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.477616 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ceph\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.477987 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.478144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmtnf\" (UniqueName: \"kubernetes.io/projected/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-kube-api-access-cmtnf\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.478401 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.478503 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.478586 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.478676 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-inventory\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.478757 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.478897 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ssh-key\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.479030 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.479164 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.479264 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.483046 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.483175 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ceph\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.483452 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.483930 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.484794 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-inventory\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.484865 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.485265 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.485506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ssh-key\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.485566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.485858 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.496118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.496157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmtnf\" (UniqueName: \"kubernetes.io/projected/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-kube-api-access-cmtnf\") pod \"install-certs-openstack-openstack-cell1-l9x7s\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:10 crc kubenswrapper[4809]: I1205 13:15:10.617119 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:11 crc kubenswrapper[4809]: I1205 13:15:11.176058 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-l9x7s"] Dec 05 13:15:11 crc kubenswrapper[4809]: W1205 13:15:11.182187 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7d9b2ef_dbdf_46f5_8618_49560290b2c4.slice/crio-297d613d04f7594570aeeb126a3e440b589d99ede78a9adb7444cf8e361626bd WatchSource:0}: Error finding container 297d613d04f7594570aeeb126a3e440b589d99ede78a9adb7444cf8e361626bd: Status 404 returned error can't find the container with id 297d613d04f7594570aeeb126a3e440b589d99ede78a9adb7444cf8e361626bd Dec 05 13:15:11 crc kubenswrapper[4809]: I1205 13:15:11.239609 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" event={"ID":"e7d9b2ef-dbdf-46f5-8618-49560290b2c4","Type":"ContainerStarted","Data":"297d613d04f7594570aeeb126a3e440b589d99ede78a9adb7444cf8e361626bd"} Dec 05 13:15:13 crc kubenswrapper[4809]: I1205 13:15:13.261707 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" event={"ID":"e7d9b2ef-dbdf-46f5-8618-49560290b2c4","Type":"ContainerStarted","Data":"9d49d91cd8e67427b6c0159a8caed67d7204d10d75a3f44ac8e472b353b20543"} Dec 05 13:15:13 crc kubenswrapper[4809]: I1205 13:15:13.286756 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" podStartSLOduration=2.340501798 podStartE2EDuration="3.286739928s" podCreationTimestamp="2025-12-05 13:15:10 +0000 UTC" firstStartedPulling="2025-12-05 13:15:11.185127125 +0000 UTC m=+7606.576103683" lastFinishedPulling="2025-12-05 13:15:12.131365255 +0000 UTC m=+7607.522341813" observedRunningTime="2025-12-05 13:15:13.282031871 +0000 UTC m=+7608.673008449" watchObservedRunningTime="2025-12-05 13:15:13.286739928 +0000 UTC m=+7608.677716486" Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.046540 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.046859 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.046899 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.047813 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1ad1f7dac7d9df8e8ac1f4d579d4e9cb79c07e17a81b299ada3af9bcee2479c"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.047876 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://a1ad1f7dac7d9df8e8ac1f4d579d4e9cb79c07e17a81b299ada3af9bcee2479c" gracePeriod=600 Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.275267 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="a1ad1f7dac7d9df8e8ac1f4d579d4e9cb79c07e17a81b299ada3af9bcee2479c" exitCode=0 Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.275400 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"a1ad1f7dac7d9df8e8ac1f4d579d4e9cb79c07e17a81b299ada3af9bcee2479c"} Dec 05 13:15:14 crc kubenswrapper[4809]: I1205 13:15:14.275474 4809 scope.go:117] "RemoveContainer" containerID="8ea43987ac79a2813b803bf54f8228234dfaa5fbd0391eedf472a8775cb0cb80" Dec 05 13:15:15 crc kubenswrapper[4809]: I1205 13:15:15.287248 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f"} Dec 05 13:15:32 crc kubenswrapper[4809]: I1205 13:15:32.477982 4809 generic.go:334] "Generic (PLEG): container finished" podID="e7d9b2ef-dbdf-46f5-8618-49560290b2c4" containerID="9d49d91cd8e67427b6c0159a8caed67d7204d10d75a3f44ac8e472b353b20543" exitCode=0 Dec 05 13:15:32 crc kubenswrapper[4809]: I1205 13:15:32.478029 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" event={"ID":"e7d9b2ef-dbdf-46f5-8618-49560290b2c4","Type":"ContainerDied","Data":"9d49d91cd8e67427b6c0159a8caed67d7204d10d75a3f44ac8e472b353b20543"} Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.052994 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.201876 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ssh-key\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202090 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-telemetry-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202169 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-bootstrap-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202311 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-dhcp-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202352 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-libvirt-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202424 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-sriov-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202485 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-metadata-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202536 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-inventory\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202625 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-nova-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202735 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmtnf\" (UniqueName: \"kubernetes.io/projected/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-kube-api-access-cmtnf\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202847 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ovn-combined-ca-bundle\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.202925 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ceph\") pod \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\" (UID: \"e7d9b2ef-dbdf-46f5-8618-49560290b2c4\") " Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.209563 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.211169 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.211481 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.212298 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.214278 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ceph" (OuterVolumeSpecName: "ceph") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.214374 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.214559 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.214958 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.216294 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.217160 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-kube-api-access-cmtnf" (OuterVolumeSpecName: "kube-api-access-cmtnf") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "kube-api-access-cmtnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.246280 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-inventory" (OuterVolumeSpecName: "inventory") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.248707 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e7d9b2ef-dbdf-46f5-8618-49560290b2c4" (UID: "e7d9b2ef-dbdf-46f5-8618-49560290b2c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306203 4809 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306255 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmtnf\" (UniqueName: \"kubernetes.io/projected/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-kube-api-access-cmtnf\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306268 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306280 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306292 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306309 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306323 4809 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306336 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306350 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306362 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306373 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.306385 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7d9b2ef-dbdf-46f5-8618-49560290b2c4-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.535837 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" event={"ID":"e7d9b2ef-dbdf-46f5-8618-49560290b2c4","Type":"ContainerDied","Data":"297d613d04f7594570aeeb126a3e440b589d99ede78a9adb7444cf8e361626bd"} Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.536148 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="297d613d04f7594570aeeb126a3e440b589d99ede78a9adb7444cf8e361626bd" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.535939 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-l9x7s" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.583106 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-dn498"] Dec 05 13:15:34 crc kubenswrapper[4809]: E1205 13:15:34.583826 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7d9b2ef-dbdf-46f5-8618-49560290b2c4" containerName="install-certs-openstack-openstack-cell1" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.583845 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7d9b2ef-dbdf-46f5-8618-49560290b2c4" containerName="install-certs-openstack-openstack-cell1" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.584086 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7d9b2ef-dbdf-46f5-8618-49560290b2c4" containerName="install-certs-openstack-openstack-cell1" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.585016 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.588806 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.589104 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.589324 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.589657 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.606868 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-dn498"] Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.715503 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.715861 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm64s\" (UniqueName: \"kubernetes.io/projected/452f44ca-1d18-4f53-a39e-74b0b4d9c942-kube-api-access-wm64s\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.715965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ceph\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.716111 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-inventory\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.817530 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm64s\" (UniqueName: \"kubernetes.io/projected/452f44ca-1d18-4f53-a39e-74b0b4d9c942-kube-api-access-wm64s\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.817579 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ceph\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.817682 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-inventory\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.817711 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.823099 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.823310 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-inventory\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.824089 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ceph\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.839271 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm64s\" (UniqueName: \"kubernetes.io/projected/452f44ca-1d18-4f53-a39e-74b0b4d9c942-kube-api-access-wm64s\") pod \"ceph-client-openstack-openstack-cell1-dn498\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:34 crc kubenswrapper[4809]: I1205 13:15:34.907180 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:35 crc kubenswrapper[4809]: I1205 13:15:35.548932 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-dn498"] Dec 05 13:15:35 crc kubenswrapper[4809]: W1205 13:15:35.555387 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod452f44ca_1d18_4f53_a39e_74b0b4d9c942.slice/crio-baa1aa7fcf8050eebfdcaf243243afef6e2ada3ca6ddf39a4c68e5692080bb85 WatchSource:0}: Error finding container baa1aa7fcf8050eebfdcaf243243afef6e2ada3ca6ddf39a4c68e5692080bb85: Status 404 returned error can't find the container with id baa1aa7fcf8050eebfdcaf243243afef6e2ada3ca6ddf39a4c68e5692080bb85 Dec 05 13:15:36 crc kubenswrapper[4809]: I1205 13:15:36.567658 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" event={"ID":"452f44ca-1d18-4f53-a39e-74b0b4d9c942","Type":"ContainerStarted","Data":"30fe54dd0b208461d29f34b24292b298a856b6276c6628ab1cdfc91d8dadc731"} Dec 05 13:15:36 crc kubenswrapper[4809]: I1205 13:15:36.569916 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" event={"ID":"452f44ca-1d18-4f53-a39e-74b0b4d9c942","Type":"ContainerStarted","Data":"baa1aa7fcf8050eebfdcaf243243afef6e2ada3ca6ddf39a4c68e5692080bb85"} Dec 05 13:15:36 crc kubenswrapper[4809]: I1205 13:15:36.610163 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" podStartSLOduration=2.088445485 podStartE2EDuration="2.610136348s" podCreationTimestamp="2025-12-05 13:15:34 +0000 UTC" firstStartedPulling="2025-12-05 13:15:35.557596595 +0000 UTC m=+7630.948573163" lastFinishedPulling="2025-12-05 13:15:36.079287458 +0000 UTC m=+7631.470264026" observedRunningTime="2025-12-05 13:15:36.587528399 +0000 UTC m=+7631.978504987" watchObservedRunningTime="2025-12-05 13:15:36.610136348 +0000 UTC m=+7632.001112946" Dec 05 13:15:41 crc kubenswrapper[4809]: I1205 13:15:41.622225 4809 generic.go:334] "Generic (PLEG): container finished" podID="452f44ca-1d18-4f53-a39e-74b0b4d9c942" containerID="30fe54dd0b208461d29f34b24292b298a856b6276c6628ab1cdfc91d8dadc731" exitCode=0 Dec 05 13:15:41 crc kubenswrapper[4809]: I1205 13:15:41.622347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" event={"ID":"452f44ca-1d18-4f53-a39e-74b0b4d9c942","Type":"ContainerDied","Data":"30fe54dd0b208461d29f34b24292b298a856b6276c6628ab1cdfc91d8dadc731"} Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.222328 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.412332 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ssh-key\") pod \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.412444 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-inventory\") pod \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.412497 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm64s\" (UniqueName: \"kubernetes.io/projected/452f44ca-1d18-4f53-a39e-74b0b4d9c942-kube-api-access-wm64s\") pod \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.412574 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ceph\") pod \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\" (UID: \"452f44ca-1d18-4f53-a39e-74b0b4d9c942\") " Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.419369 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452f44ca-1d18-4f53-a39e-74b0b4d9c942-kube-api-access-wm64s" (OuterVolumeSpecName: "kube-api-access-wm64s") pod "452f44ca-1d18-4f53-a39e-74b0b4d9c942" (UID: "452f44ca-1d18-4f53-a39e-74b0b4d9c942"). InnerVolumeSpecName "kube-api-access-wm64s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.445786 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ceph" (OuterVolumeSpecName: "ceph") pod "452f44ca-1d18-4f53-a39e-74b0b4d9c942" (UID: "452f44ca-1d18-4f53-a39e-74b0b4d9c942"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.515155 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm64s\" (UniqueName: \"kubernetes.io/projected/452f44ca-1d18-4f53-a39e-74b0b4d9c942-kube-api-access-wm64s\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.515208 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.515628 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-inventory" (OuterVolumeSpecName: "inventory") pod "452f44ca-1d18-4f53-a39e-74b0b4d9c942" (UID: "452f44ca-1d18-4f53-a39e-74b0b4d9c942"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.524813 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "452f44ca-1d18-4f53-a39e-74b0b4d9c942" (UID: "452f44ca-1d18-4f53-a39e-74b0b4d9c942"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.618036 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.618074 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/452f44ca-1d18-4f53-a39e-74b0b4d9c942-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.644208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" event={"ID":"452f44ca-1d18-4f53-a39e-74b0b4d9c942","Type":"ContainerDied","Data":"baa1aa7fcf8050eebfdcaf243243afef6e2ada3ca6ddf39a4c68e5692080bb85"} Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.644255 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baa1aa7fcf8050eebfdcaf243243afef6e2ada3ca6ddf39a4c68e5692080bb85" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.644252 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-dn498" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.741937 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rvkzx"] Dec 05 13:15:43 crc kubenswrapper[4809]: E1205 13:15:43.742353 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452f44ca-1d18-4f53-a39e-74b0b4d9c942" containerName="ceph-client-openstack-openstack-cell1" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.742368 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="452f44ca-1d18-4f53-a39e-74b0b4d9c942" containerName="ceph-client-openstack-openstack-cell1" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.742575 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="452f44ca-1d18-4f53-a39e-74b0b4d9c942" containerName="ceph-client-openstack-openstack-cell1" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.743307 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.745185 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.745492 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.745597 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.745719 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.747344 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.756051 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rvkzx"] Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.823103 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-inventory\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.823211 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/db861e9a-684e-4ca6-99b6-c07b45229909-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.823468 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.823593 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ceph\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.823664 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ssh-key\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.823688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5nfx\" (UniqueName: \"kubernetes.io/projected/db861e9a-684e-4ca6-99b6-c07b45229909-kube-api-access-t5nfx\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.925835 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5nfx\" (UniqueName: \"kubernetes.io/projected/db861e9a-684e-4ca6-99b6-c07b45229909-kube-api-access-t5nfx\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.925950 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-inventory\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.926001 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/db861e9a-684e-4ca6-99b6-c07b45229909-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.926124 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.926200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ceph\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.926234 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ssh-key\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.927785 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/db861e9a-684e-4ca6-99b6-c07b45229909-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.930499 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ssh-key\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.930747 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ceph\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.931148 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-inventory\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.933076 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:43 crc kubenswrapper[4809]: I1205 13:15:43.945508 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5nfx\" (UniqueName: \"kubernetes.io/projected/db861e9a-684e-4ca6-99b6-c07b45229909-kube-api-access-t5nfx\") pod \"ovn-openstack-openstack-cell1-rvkzx\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:44 crc kubenswrapper[4809]: I1205 13:15:44.062184 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:15:44 crc kubenswrapper[4809]: I1205 13:15:44.609147 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rvkzx"] Dec 05 13:15:44 crc kubenswrapper[4809]: I1205 13:15:44.655383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" event={"ID":"db861e9a-684e-4ca6-99b6-c07b45229909","Type":"ContainerStarted","Data":"a3cde3089232526fec8867a539fe51abe0b6ba85eec63f067f6a6fab44441ccb"} Dec 05 13:15:45 crc kubenswrapper[4809]: I1205 13:15:45.670127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" event={"ID":"db861e9a-684e-4ca6-99b6-c07b45229909","Type":"ContainerStarted","Data":"f7ed8483ea76d91e4a3bdb834356f90bc15c085bd3708b2c73a9c221fca28852"} Dec 05 13:15:45 crc kubenswrapper[4809]: I1205 13:15:45.695482 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" podStartSLOduration=2.111845385 podStartE2EDuration="2.695455727s" podCreationTimestamp="2025-12-05 13:15:43 +0000 UTC" firstStartedPulling="2025-12-05 13:15:44.611003564 +0000 UTC m=+7640.001980122" lastFinishedPulling="2025-12-05 13:15:45.194613906 +0000 UTC m=+7640.585590464" observedRunningTime="2025-12-05 13:15:45.687732239 +0000 UTC m=+7641.078708807" watchObservedRunningTime="2025-12-05 13:15:45.695455727 +0000 UTC m=+7641.086432285" Dec 05 13:15:46 crc kubenswrapper[4809]: I1205 13:15:46.599051 4809 scope.go:117] "RemoveContainer" containerID="12b41af7869365eb8f34cb87c168476d6ac82f5d8b614b8ad6a94b91969e2931" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.582480 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z6vfh"] Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.586136 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.597013 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6vfh"] Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.757812 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-catalog-content\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.757925 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5pfk\" (UniqueName: \"kubernetes.io/projected/2368ec1d-c426-4720-a767-517aa1ac5e5a-kube-api-access-s5pfk\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.758035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-utilities\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.860117 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-catalog-content\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.860185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5pfk\" (UniqueName: \"kubernetes.io/projected/2368ec1d-c426-4720-a767-517aa1ac5e5a-kube-api-access-s5pfk\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.860235 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-utilities\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.860779 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-catalog-content\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.860805 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-utilities\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.885669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5pfk\" (UniqueName: \"kubernetes.io/projected/2368ec1d-c426-4720-a767-517aa1ac5e5a-kube-api-access-s5pfk\") pod \"redhat-operators-z6vfh\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:19 crc kubenswrapper[4809]: I1205 13:16:19.915666 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:20 crc kubenswrapper[4809]: I1205 13:16:20.424749 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6vfh"] Dec 05 13:16:21 crc kubenswrapper[4809]: I1205 13:16:21.143609 4809 generic.go:334] "Generic (PLEG): container finished" podID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerID="ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445" exitCode=0 Dec 05 13:16:21 crc kubenswrapper[4809]: I1205 13:16:21.143813 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6vfh" event={"ID":"2368ec1d-c426-4720-a767-517aa1ac5e5a","Type":"ContainerDied","Data":"ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445"} Dec 05 13:16:21 crc kubenswrapper[4809]: I1205 13:16:21.143930 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6vfh" event={"ID":"2368ec1d-c426-4720-a767-517aa1ac5e5a","Type":"ContainerStarted","Data":"348c713cab09a97244b1f75048adefab8a6dc4ff487475924e43f055b3cfc0b0"} Dec 05 13:16:22 crc kubenswrapper[4809]: I1205 13:16:22.155605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6vfh" event={"ID":"2368ec1d-c426-4720-a767-517aa1ac5e5a","Type":"ContainerStarted","Data":"bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5"} Dec 05 13:16:26 crc kubenswrapper[4809]: I1205 13:16:26.198105 4809 generic.go:334] "Generic (PLEG): container finished" podID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerID="bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5" exitCode=0 Dec 05 13:16:26 crc kubenswrapper[4809]: I1205 13:16:26.198196 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6vfh" event={"ID":"2368ec1d-c426-4720-a767-517aa1ac5e5a","Type":"ContainerDied","Data":"bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5"} Dec 05 13:16:27 crc kubenswrapper[4809]: I1205 13:16:27.212200 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6vfh" event={"ID":"2368ec1d-c426-4720-a767-517aa1ac5e5a","Type":"ContainerStarted","Data":"4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28"} Dec 05 13:16:27 crc kubenswrapper[4809]: I1205 13:16:27.238450 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z6vfh" podStartSLOduration=2.769973176 podStartE2EDuration="8.238426073s" podCreationTimestamp="2025-12-05 13:16:19 +0000 UTC" firstStartedPulling="2025-12-05 13:16:21.146091469 +0000 UTC m=+7676.537068017" lastFinishedPulling="2025-12-05 13:16:26.614544356 +0000 UTC m=+7682.005520914" observedRunningTime="2025-12-05 13:16:27.231121986 +0000 UTC m=+7682.622098554" watchObservedRunningTime="2025-12-05 13:16:27.238426073 +0000 UTC m=+7682.629402631" Dec 05 13:16:29 crc kubenswrapper[4809]: I1205 13:16:29.917134 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:29 crc kubenswrapper[4809]: I1205 13:16:29.918101 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:30 crc kubenswrapper[4809]: I1205 13:16:30.989977 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z6vfh" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="registry-server" probeResult="failure" output=< Dec 05 13:16:30 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 13:16:30 crc kubenswrapper[4809]: > Dec 05 13:16:38 crc kubenswrapper[4809]: I1205 13:16:38.336973 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-pgs44" podUID="864e3e5e-8812-4482-8cf0-a36c2e3c4d36" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 13:16:40 crc kubenswrapper[4809]: I1205 13:16:40.966971 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z6vfh" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="registry-server" probeResult="failure" output=< Dec 05 13:16:40 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 13:16:40 crc kubenswrapper[4809]: > Dec 05 13:16:49 crc kubenswrapper[4809]: I1205 13:16:49.978215 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:50 crc kubenswrapper[4809]: I1205 13:16:50.031913 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:50 crc kubenswrapper[4809]: I1205 13:16:50.789287 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6vfh"] Dec 05 13:16:51 crc kubenswrapper[4809]: I1205 13:16:51.508229 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z6vfh" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="registry-server" containerID="cri-o://4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28" gracePeriod=2 Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.506251 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.527390 4809 generic.go:334] "Generic (PLEG): container finished" podID="db861e9a-684e-4ca6-99b6-c07b45229909" containerID="f7ed8483ea76d91e4a3bdb834356f90bc15c085bd3708b2c73a9c221fca28852" exitCode=0 Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.527489 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" event={"ID":"db861e9a-684e-4ca6-99b6-c07b45229909","Type":"ContainerDied","Data":"f7ed8483ea76d91e4a3bdb834356f90bc15c085bd3708b2c73a9c221fca28852"} Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.540161 4809 generic.go:334] "Generic (PLEG): container finished" podID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerID="4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28" exitCode=0 Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.540315 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6vfh" event={"ID":"2368ec1d-c426-4720-a767-517aa1ac5e5a","Type":"ContainerDied","Data":"4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28"} Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.540459 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6vfh" event={"ID":"2368ec1d-c426-4720-a767-517aa1ac5e5a","Type":"ContainerDied","Data":"348c713cab09a97244b1f75048adefab8a6dc4ff487475924e43f055b3cfc0b0"} Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.540490 4809 scope.go:117] "RemoveContainer" containerID="4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.541307 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6vfh" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.580924 4809 scope.go:117] "RemoveContainer" containerID="bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.612808 4809 scope.go:117] "RemoveContainer" containerID="ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.663094 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-utilities\") pod \"2368ec1d-c426-4720-a767-517aa1ac5e5a\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.663231 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5pfk\" (UniqueName: \"kubernetes.io/projected/2368ec1d-c426-4720-a767-517aa1ac5e5a-kube-api-access-s5pfk\") pod \"2368ec1d-c426-4720-a767-517aa1ac5e5a\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.663372 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-catalog-content\") pod \"2368ec1d-c426-4720-a767-517aa1ac5e5a\" (UID: \"2368ec1d-c426-4720-a767-517aa1ac5e5a\") " Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.664858 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-utilities" (OuterVolumeSpecName: "utilities") pod "2368ec1d-c426-4720-a767-517aa1ac5e5a" (UID: "2368ec1d-c426-4720-a767-517aa1ac5e5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.669550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2368ec1d-c426-4720-a767-517aa1ac5e5a-kube-api-access-s5pfk" (OuterVolumeSpecName: "kube-api-access-s5pfk") pod "2368ec1d-c426-4720-a767-517aa1ac5e5a" (UID: "2368ec1d-c426-4720-a767-517aa1ac5e5a"). InnerVolumeSpecName "kube-api-access-s5pfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.672164 4809 scope.go:117] "RemoveContainer" containerID="4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28" Dec 05 13:16:52 crc kubenswrapper[4809]: E1205 13:16:52.673667 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28\": container with ID starting with 4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28 not found: ID does not exist" containerID="4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.673723 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28"} err="failed to get container status \"4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28\": rpc error: code = NotFound desc = could not find container \"4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28\": container with ID starting with 4467b8165b766884d52d3a83ddc48f539ea96bfa43e9415fd2fd88902882ed28 not found: ID does not exist" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.673759 4809 scope.go:117] "RemoveContainer" containerID="bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5" Dec 05 13:16:52 crc kubenswrapper[4809]: E1205 13:16:52.674190 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5\": container with ID starting with bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5 not found: ID does not exist" containerID="bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.674220 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5"} err="failed to get container status \"bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5\": rpc error: code = NotFound desc = could not find container \"bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5\": container with ID starting with bb7d93558db7e6fd8fc4d3b70d448c9ffad8dfb959454443b8a099eaaafc76c5 not found: ID does not exist" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.674237 4809 scope.go:117] "RemoveContainer" containerID="ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445" Dec 05 13:16:52 crc kubenswrapper[4809]: E1205 13:16:52.675670 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445\": container with ID starting with ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445 not found: ID does not exist" containerID="ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.675718 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445"} err="failed to get container status \"ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445\": rpc error: code = NotFound desc = could not find container \"ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445\": container with ID starting with ed5a34717dd57843b644e8081f651aa814c735d97be58dc134961883813c5445 not found: ID does not exist" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.766821 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5pfk\" (UniqueName: \"kubernetes.io/projected/2368ec1d-c426-4720-a767-517aa1ac5e5a-kube-api-access-s5pfk\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.766859 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.774222 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2368ec1d-c426-4720-a767-517aa1ac5e5a" (UID: "2368ec1d-c426-4720-a767-517aa1ac5e5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.872624 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2368ec1d-c426-4720-a767-517aa1ac5e5a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.892646 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6vfh"] Dec 05 13:16:52 crc kubenswrapper[4809]: I1205 13:16:52.897002 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z6vfh"] Dec 05 13:16:53 crc kubenswrapper[4809]: I1205 13:16:53.993451 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.099273 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5nfx\" (UniqueName: \"kubernetes.io/projected/db861e9a-684e-4ca6-99b6-c07b45229909-kube-api-access-t5nfx\") pod \"db861e9a-684e-4ca6-99b6-c07b45229909\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.099467 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ovn-combined-ca-bundle\") pod \"db861e9a-684e-4ca6-99b6-c07b45229909\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.099492 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ssh-key\") pod \"db861e9a-684e-4ca6-99b6-c07b45229909\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.099553 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/db861e9a-684e-4ca6-99b6-c07b45229909-ovncontroller-config-0\") pod \"db861e9a-684e-4ca6-99b6-c07b45229909\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.099583 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ceph\") pod \"db861e9a-684e-4ca6-99b6-c07b45229909\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.099650 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-inventory\") pod \"db861e9a-684e-4ca6-99b6-c07b45229909\" (UID: \"db861e9a-684e-4ca6-99b6-c07b45229909\") " Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.104970 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ceph" (OuterVolumeSpecName: "ceph") pod "db861e9a-684e-4ca6-99b6-c07b45229909" (UID: "db861e9a-684e-4ca6-99b6-c07b45229909"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.105164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "db861e9a-684e-4ca6-99b6-c07b45229909" (UID: "db861e9a-684e-4ca6-99b6-c07b45229909"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.113197 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db861e9a-684e-4ca6-99b6-c07b45229909-kube-api-access-t5nfx" (OuterVolumeSpecName: "kube-api-access-t5nfx") pod "db861e9a-684e-4ca6-99b6-c07b45229909" (UID: "db861e9a-684e-4ca6-99b6-c07b45229909"). InnerVolumeSpecName "kube-api-access-t5nfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.134020 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db861e9a-684e-4ca6-99b6-c07b45229909" (UID: "db861e9a-684e-4ca6-99b6-c07b45229909"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.141718 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db861e9a-684e-4ca6-99b6-c07b45229909-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "db861e9a-684e-4ca6-99b6-c07b45229909" (UID: "db861e9a-684e-4ca6-99b6-c07b45229909"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.145537 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-inventory" (OuterVolumeSpecName: "inventory") pod "db861e9a-684e-4ca6-99b6-c07b45229909" (UID: "db861e9a-684e-4ca6-99b6-c07b45229909"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.202218 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.202357 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.202440 4809 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/db861e9a-684e-4ca6-99b6-c07b45229909-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.202518 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.202602 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db861e9a-684e-4ca6-99b6-c07b45229909-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.202702 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5nfx\" (UniqueName: \"kubernetes.io/projected/db861e9a-684e-4ca6-99b6-c07b45229909-kube-api-access-t5nfx\") on node \"crc\" DevicePath \"\"" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.566134 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" event={"ID":"db861e9a-684e-4ca6-99b6-c07b45229909","Type":"ContainerDied","Data":"a3cde3089232526fec8867a539fe51abe0b6ba85eec63f067f6a6fab44441ccb"} Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.566194 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3cde3089232526fec8867a539fe51abe0b6ba85eec63f067f6a6fab44441ccb" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.566261 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rvkzx" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.650673 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-2cr2r"] Dec 05 13:16:54 crc kubenswrapper[4809]: E1205 13:16:54.651203 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db861e9a-684e-4ca6-99b6-c07b45229909" containerName="ovn-openstack-openstack-cell1" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.651231 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="db861e9a-684e-4ca6-99b6-c07b45229909" containerName="ovn-openstack-openstack-cell1" Dec 05 13:16:54 crc kubenswrapper[4809]: E1205 13:16:54.651266 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="extract-utilities" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.651277 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="extract-utilities" Dec 05 13:16:54 crc kubenswrapper[4809]: E1205 13:16:54.651299 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="extract-content" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.651309 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="extract-content" Dec 05 13:16:54 crc kubenswrapper[4809]: E1205 13:16:54.651327 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="registry-server" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.651334 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="registry-server" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.651579 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" containerName="registry-server" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.651615 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="db861e9a-684e-4ca6-99b6-c07b45229909" containerName="ovn-openstack-openstack-cell1" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.652588 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.654959 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.654985 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.655125 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.655224 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.655291 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.655452 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.681599 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-2cr2r"] Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.813553 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.813866 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.814163 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.814308 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.814488 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.814546 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5mk2\" (UniqueName: \"kubernetes.io/projected/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-kube-api-access-j5mk2\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.814578 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.885164 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2368ec1d-c426-4720-a767-517aa1ac5e5a" path="/var/lib/kubelet/pods/2368ec1d-c426-4720-a767-517aa1ac5e5a/volumes" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.916542 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.916622 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.916752 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.916800 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.916869 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.916904 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5mk2\" (UniqueName: \"kubernetes.io/projected/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-kube-api-access-j5mk2\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.916931 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.922408 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.922483 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.922517 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.922680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.923103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.923160 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.933868 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5mk2\" (UniqueName: \"kubernetes.io/projected/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-kube-api-access-j5mk2\") pod \"neutron-metadata-openstack-openstack-cell1-2cr2r\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:54 crc kubenswrapper[4809]: I1205 13:16:54.980207 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:16:55 crc kubenswrapper[4809]: I1205 13:16:55.534720 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-2cr2r"] Dec 05 13:16:55 crc kubenswrapper[4809]: I1205 13:16:55.585314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" event={"ID":"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7","Type":"ContainerStarted","Data":"73f75066d31e4deeec9e580769218a066cfa2eb164be5ba873911fde875c7542"} Dec 05 13:16:58 crc kubenswrapper[4809]: I1205 13:16:58.617418 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" event={"ID":"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7","Type":"ContainerStarted","Data":"e01ad8882ba6c8f0a76b6ed083f689b1c3c4aa758dd604252b1f292dfb9028f9"} Dec 05 13:16:58 crc kubenswrapper[4809]: I1205 13:16:58.642193 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" podStartSLOduration=3.141604484 podStartE2EDuration="4.642169628s" podCreationTimestamp="2025-12-05 13:16:54 +0000 UTC" firstStartedPulling="2025-12-05 13:16:55.543697091 +0000 UTC m=+7710.934673639" lastFinishedPulling="2025-12-05 13:16:57.044262225 +0000 UTC m=+7712.435238783" observedRunningTime="2025-12-05 13:16:58.632028405 +0000 UTC m=+7714.023004963" watchObservedRunningTime="2025-12-05 13:16:58.642169628 +0000 UTC m=+7714.033146186" Dec 05 13:17:14 crc kubenswrapper[4809]: I1205 13:17:14.046670 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:17:14 crc kubenswrapper[4809]: I1205 13:17:14.047162 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:17:15 crc kubenswrapper[4809]: I1205 13:17:15.520962 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-7n7zc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 13:17:15 crc kubenswrapper[4809]: I1205 13:17:15.521321 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" podUID="ddd951b5-5bbf-4f65-b09a-4482bb58fcb6" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 13:17:44 crc kubenswrapper[4809]: I1205 13:17:44.046745 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:17:44 crc kubenswrapper[4809]: I1205 13:17:44.047494 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:17:54 crc kubenswrapper[4809]: I1205 13:17:54.270780 4809 generic.go:334] "Generic (PLEG): container finished" podID="31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" containerID="e01ad8882ba6c8f0a76b6ed083f689b1c3c4aa758dd604252b1f292dfb9028f9" exitCode=0 Dec 05 13:17:54 crc kubenswrapper[4809]: I1205 13:17:54.270821 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" event={"ID":"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7","Type":"ContainerDied","Data":"e01ad8882ba6c8f0a76b6ed083f689b1c3c4aa758dd604252b1f292dfb9028f9"} Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.726875 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.757995 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.758140 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ssh-key\") pod \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.758168 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5mk2\" (UniqueName: \"kubernetes.io/projected/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-kube-api-access-j5mk2\") pod \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.758241 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-nova-metadata-neutron-config-0\") pod \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.758267 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ceph\") pod \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.758366 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-inventory\") pod \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.758403 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-metadata-combined-ca-bundle\") pod \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\" (UID: \"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7\") " Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.767229 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ceph" (OuterVolumeSpecName: "ceph") pod "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" (UID: "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.767319 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" (UID: "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.781018 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-kube-api-access-j5mk2" (OuterVolumeSpecName: "kube-api-access-j5mk2") pod "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" (UID: "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7"). InnerVolumeSpecName "kube-api-access-j5mk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.800038 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" (UID: "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.808428 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" (UID: "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.810362 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-inventory" (OuterVolumeSpecName: "inventory") pod "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" (UID: "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.822196 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" (UID: "31f3c035-3079-4cf5-ab18-b4a1db3fb3b7"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.862917 4809 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.862974 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.863003 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.863027 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.863057 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.863082 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:17:55 crc kubenswrapper[4809]: I1205 13:17:55.863106 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5mk2\" (UniqueName: \"kubernetes.io/projected/31f3c035-3079-4cf5-ab18-b4a1db3fb3b7-kube-api-access-j5mk2\") on node \"crc\" DevicePath \"\"" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.298558 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" event={"ID":"31f3c035-3079-4cf5-ab18-b4a1db3fb3b7","Type":"ContainerDied","Data":"73f75066d31e4deeec9e580769218a066cfa2eb164be5ba873911fde875c7542"} Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.298606 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73f75066d31e4deeec9e580769218a066cfa2eb164be5ba873911fde875c7542" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.298696 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2cr2r" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.438404 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-m8cjr"] Dec 05 13:17:56 crc kubenswrapper[4809]: E1205 13:17:56.438984 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" containerName="neutron-metadata-openstack-openstack-cell1" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.439007 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" containerName="neutron-metadata-openstack-openstack-cell1" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.439296 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="31f3c035-3079-4cf5-ab18-b4a1db3fb3b7" containerName="neutron-metadata-openstack-openstack-cell1" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.440210 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.443405 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.443810 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.443904 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.445124 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.445394 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.449441 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-m8cjr"] Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.476835 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.476924 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.476965 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ssh-key\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.477105 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-inventory\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.477268 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ceph\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.477313 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4ztn\" (UniqueName: \"kubernetes.io/projected/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-kube-api-access-w4ztn\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.579200 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.579284 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ssh-key\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.579396 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-inventory\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.579501 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ceph\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.579545 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4ztn\" (UniqueName: \"kubernetes.io/projected/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-kube-api-access-w4ztn\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.579705 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.586698 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ceph\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.587182 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ssh-key\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.587744 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.588021 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-inventory\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.588992 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.612481 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4ztn\" (UniqueName: \"kubernetes.io/projected/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-kube-api-access-w4ztn\") pod \"libvirt-openstack-openstack-cell1-m8cjr\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:56 crc kubenswrapper[4809]: I1205 13:17:56.767860 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:17:57 crc kubenswrapper[4809]: I1205 13:17:57.424325 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-m8cjr"] Dec 05 13:17:57 crc kubenswrapper[4809]: W1205 13:17:57.461573 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod328ce8d0_f42c_4e66_9de5_0f0cac6a2994.slice/crio-c81ce1356c8952310f0bac91eac73e33597443783f58c5e52577762c1cc19f6a WatchSource:0}: Error finding container c81ce1356c8952310f0bac91eac73e33597443783f58c5e52577762c1cc19f6a: Status 404 returned error can't find the container with id c81ce1356c8952310f0bac91eac73e33597443783f58c5e52577762c1cc19f6a Dec 05 13:17:58 crc kubenswrapper[4809]: I1205 13:17:58.321288 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" event={"ID":"328ce8d0-f42c-4e66-9de5-0f0cac6a2994","Type":"ContainerStarted","Data":"c81ce1356c8952310f0bac91eac73e33597443783f58c5e52577762c1cc19f6a"} Dec 05 13:17:59 crc kubenswrapper[4809]: I1205 13:17:59.337915 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" event={"ID":"328ce8d0-f42c-4e66-9de5-0f0cac6a2994","Type":"ContainerStarted","Data":"7548cc8b1a61d8e86f7e96ce1ef1e6b1970b18df61740eb1551ff22d1804ac8d"} Dec 05 13:17:59 crc kubenswrapper[4809]: I1205 13:17:59.358109 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" podStartSLOduration=2.556264728 podStartE2EDuration="3.358082456s" podCreationTimestamp="2025-12-05 13:17:56 +0000 UTC" firstStartedPulling="2025-12-05 13:17:57.463842122 +0000 UTC m=+7772.854818690" lastFinishedPulling="2025-12-05 13:17:58.26565985 +0000 UTC m=+7773.656636418" observedRunningTime="2025-12-05 13:17:59.353236896 +0000 UTC m=+7774.744213474" watchObservedRunningTime="2025-12-05 13:17:59.358082456 +0000 UTC m=+7774.749059014" Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.046600 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.047468 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.047539 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.048554 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.048605 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" gracePeriod=600 Dec 05 13:18:14 crc kubenswrapper[4809]: E1205 13:18:14.173600 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.504741 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" exitCode=0 Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.504781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f"} Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.504819 4809 scope.go:117] "RemoveContainer" containerID="a1ad1f7dac7d9df8e8ac1f4d579d4e9cb79c07e17a81b299ada3af9bcee2479c" Dec 05 13:18:14 crc kubenswrapper[4809]: I1205 13:18:14.505416 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:18:14 crc kubenswrapper[4809]: E1205 13:18:14.505694 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:18:25 crc kubenswrapper[4809]: I1205 13:18:25.872133 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:18:25 crc kubenswrapper[4809]: E1205 13:18:25.873082 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:18:38 crc kubenswrapper[4809]: I1205 13:18:38.874339 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:18:38 crc kubenswrapper[4809]: E1205 13:18:38.875121 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:18:49 crc kubenswrapper[4809]: I1205 13:18:49.871906 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:18:49 crc kubenswrapper[4809]: E1205 13:18:49.872475 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:19:01 crc kubenswrapper[4809]: I1205 13:19:01.874206 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:19:01 crc kubenswrapper[4809]: E1205 13:19:01.875730 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:19:15 crc kubenswrapper[4809]: I1205 13:19:15.876001 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:19:15 crc kubenswrapper[4809]: E1205 13:19:15.878654 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:19:27 crc kubenswrapper[4809]: I1205 13:19:27.874144 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:19:27 crc kubenswrapper[4809]: E1205 13:19:27.875249 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:19:40 crc kubenswrapper[4809]: I1205 13:19:40.873288 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:19:40 crc kubenswrapper[4809]: E1205 13:19:40.874345 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:19:53 crc kubenswrapper[4809]: I1205 13:19:53.873108 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:19:53 crc kubenswrapper[4809]: E1205 13:19:53.873944 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:20:06 crc kubenswrapper[4809]: I1205 13:20:06.872833 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:20:06 crc kubenswrapper[4809]: E1205 13:20:06.873654 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:20:21 crc kubenswrapper[4809]: I1205 13:20:21.872476 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:20:21 crc kubenswrapper[4809]: E1205 13:20:21.873927 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:20:34 crc kubenswrapper[4809]: I1205 13:20:34.884687 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:20:34 crc kubenswrapper[4809]: E1205 13:20:34.885554 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:20:45 crc kubenswrapper[4809]: I1205 13:20:45.873140 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:20:45 crc kubenswrapper[4809]: E1205 13:20:45.873938 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.156582 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xn464"] Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.160592 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.181591 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xn464"] Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.354111 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-utilities\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.354206 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bsms\" (UniqueName: \"kubernetes.io/projected/81f4a69a-33ba-4652-8962-9179da9eb3d7-kube-api-access-6bsms\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.354246 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-catalog-content\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.456987 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-utilities\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.457139 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bsms\" (UniqueName: \"kubernetes.io/projected/81f4a69a-33ba-4652-8962-9179da9eb3d7-kube-api-access-6bsms\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.457202 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-catalog-content\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.457918 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-catalog-content\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.458183 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-utilities\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.477679 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bsms\" (UniqueName: \"kubernetes.io/projected/81f4a69a-33ba-4652-8962-9179da9eb3d7-kube-api-access-6bsms\") pod \"certified-operators-xn464\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:00 crc kubenswrapper[4809]: I1205 13:21:00.495204 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:01 crc kubenswrapper[4809]: I1205 13:21:00.875163 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:21:01 crc kubenswrapper[4809]: E1205 13:21:00.875707 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:21:02 crc kubenswrapper[4809]: I1205 13:21:02.164326 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xn464"] Dec 05 13:21:02 crc kubenswrapper[4809]: I1205 13:21:02.588650 4809 generic.go:334] "Generic (PLEG): container finished" podID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerID="acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a" exitCode=0 Dec 05 13:21:02 crc kubenswrapper[4809]: I1205 13:21:02.589423 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xn464" event={"ID":"81f4a69a-33ba-4652-8962-9179da9eb3d7","Type":"ContainerDied","Data":"acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a"} Dec 05 13:21:02 crc kubenswrapper[4809]: I1205 13:21:02.589485 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xn464" event={"ID":"81f4a69a-33ba-4652-8962-9179da9eb3d7","Type":"ContainerStarted","Data":"2a06a95477d7a6971cd76d885d9b8059a1ad15aa76bd3a98d8ccd51d1e5b3708"} Dec 05 13:21:02 crc kubenswrapper[4809]: I1205 13:21:02.592268 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:21:04 crc kubenswrapper[4809]: I1205 13:21:04.613664 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xn464" event={"ID":"81f4a69a-33ba-4652-8962-9179da9eb3d7","Type":"ContainerStarted","Data":"6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3"} Dec 05 13:21:05 crc kubenswrapper[4809]: I1205 13:21:05.629070 4809 generic.go:334] "Generic (PLEG): container finished" podID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerID="6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3" exitCode=0 Dec 05 13:21:05 crc kubenswrapper[4809]: I1205 13:21:05.629158 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xn464" event={"ID":"81f4a69a-33ba-4652-8962-9179da9eb3d7","Type":"ContainerDied","Data":"6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3"} Dec 05 13:21:06 crc kubenswrapper[4809]: I1205 13:21:06.641781 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xn464" event={"ID":"81f4a69a-33ba-4652-8962-9179da9eb3d7","Type":"ContainerStarted","Data":"209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76"} Dec 05 13:21:10 crc kubenswrapper[4809]: I1205 13:21:10.496468 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:10 crc kubenswrapper[4809]: I1205 13:21:10.497057 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:10 crc kubenswrapper[4809]: I1205 13:21:10.554669 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:10 crc kubenswrapper[4809]: I1205 13:21:10.577648 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xn464" podStartSLOduration=6.900149131 podStartE2EDuration="10.577610993s" podCreationTimestamp="2025-12-05 13:21:00 +0000 UTC" firstStartedPulling="2025-12-05 13:21:02.591825649 +0000 UTC m=+7957.982802207" lastFinishedPulling="2025-12-05 13:21:06.269287501 +0000 UTC m=+7961.660264069" observedRunningTime="2025-12-05 13:21:06.681945212 +0000 UTC m=+7962.072921780" watchObservedRunningTime="2025-12-05 13:21:10.577610993 +0000 UTC m=+7965.968587551" Dec 05 13:21:13 crc kubenswrapper[4809]: I1205 13:21:13.872546 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:21:13 crc kubenswrapper[4809]: E1205 13:21:13.873287 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:21:20 crc kubenswrapper[4809]: I1205 13:21:20.562519 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:20 crc kubenswrapper[4809]: I1205 13:21:20.641919 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xn464"] Dec 05 13:21:20 crc kubenswrapper[4809]: I1205 13:21:20.818268 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xn464" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="registry-server" containerID="cri-o://209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76" gracePeriod=2 Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.485294 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.611045 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bsms\" (UniqueName: \"kubernetes.io/projected/81f4a69a-33ba-4652-8962-9179da9eb3d7-kube-api-access-6bsms\") pod \"81f4a69a-33ba-4652-8962-9179da9eb3d7\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.611885 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-catalog-content\") pod \"81f4a69a-33ba-4652-8962-9179da9eb3d7\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.612015 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-utilities\") pod \"81f4a69a-33ba-4652-8962-9179da9eb3d7\" (UID: \"81f4a69a-33ba-4652-8962-9179da9eb3d7\") " Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.613041 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-utilities" (OuterVolumeSpecName: "utilities") pod "81f4a69a-33ba-4652-8962-9179da9eb3d7" (UID: "81f4a69a-33ba-4652-8962-9179da9eb3d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.613328 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.618112 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f4a69a-33ba-4652-8962-9179da9eb3d7-kube-api-access-6bsms" (OuterVolumeSpecName: "kube-api-access-6bsms") pod "81f4a69a-33ba-4652-8962-9179da9eb3d7" (UID: "81f4a69a-33ba-4652-8962-9179da9eb3d7"). InnerVolumeSpecName "kube-api-access-6bsms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.663347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81f4a69a-33ba-4652-8962-9179da9eb3d7" (UID: "81f4a69a-33ba-4652-8962-9179da9eb3d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.716195 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f4a69a-33ba-4652-8962-9179da9eb3d7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.716244 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bsms\" (UniqueName: \"kubernetes.io/projected/81f4a69a-33ba-4652-8962-9179da9eb3d7-kube-api-access-6bsms\") on node \"crc\" DevicePath \"\"" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.834655 4809 generic.go:334] "Generic (PLEG): container finished" podID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerID="209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76" exitCode=0 Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.834704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xn464" event={"ID":"81f4a69a-33ba-4652-8962-9179da9eb3d7","Type":"ContainerDied","Data":"209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76"} Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.834769 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xn464" event={"ID":"81f4a69a-33ba-4652-8962-9179da9eb3d7","Type":"ContainerDied","Data":"2a06a95477d7a6971cd76d885d9b8059a1ad15aa76bd3a98d8ccd51d1e5b3708"} Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.834793 4809 scope.go:117] "RemoveContainer" containerID="209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.834780 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xn464" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.865363 4809 scope.go:117] "RemoveContainer" containerID="6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.894946 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xn464"] Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.903031 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xn464"] Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.903299 4809 scope.go:117] "RemoveContainer" containerID="acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.962361 4809 scope.go:117] "RemoveContainer" containerID="209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76" Dec 05 13:21:21 crc kubenswrapper[4809]: E1205 13:21:21.962978 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76\": container with ID starting with 209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76 not found: ID does not exist" containerID="209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.963029 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76"} err="failed to get container status \"209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76\": rpc error: code = NotFound desc = could not find container \"209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76\": container with ID starting with 209c1b7e37af89f1254ee17e818823ddb98ec3d7eacabbf714350443da99da76 not found: ID does not exist" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.963056 4809 scope.go:117] "RemoveContainer" containerID="6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3" Dec 05 13:21:21 crc kubenswrapper[4809]: E1205 13:21:21.963903 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3\": container with ID starting with 6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3 not found: ID does not exist" containerID="6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.963972 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3"} err="failed to get container status \"6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3\": rpc error: code = NotFound desc = could not find container \"6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3\": container with ID starting with 6759e65ad4789edf97ff13453127d1092ea4feef3c62c30ebcaa1b7a91cc15c3 not found: ID does not exist" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.964021 4809 scope.go:117] "RemoveContainer" containerID="acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a" Dec 05 13:21:21 crc kubenswrapper[4809]: E1205 13:21:21.964714 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a\": container with ID starting with acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a not found: ID does not exist" containerID="acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a" Dec 05 13:21:21 crc kubenswrapper[4809]: I1205 13:21:21.964740 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a"} err="failed to get container status \"acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a\": rpc error: code = NotFound desc = could not find container \"acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a\": container with ID starting with acc5e787f78698cb212ec633ce23a3b9563106cd9d7ecd5ef24094d9c782ce4a not found: ID does not exist" Dec 05 13:21:22 crc kubenswrapper[4809]: I1205 13:21:22.892536 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" path="/var/lib/kubelet/pods/81f4a69a-33ba-4652-8962-9179da9eb3d7/volumes" Dec 05 13:21:26 crc kubenswrapper[4809]: I1205 13:21:26.873892 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:21:26 crc kubenswrapper[4809]: E1205 13:21:26.874914 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:21:40 crc kubenswrapper[4809]: I1205 13:21:40.874293 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:21:40 crc kubenswrapper[4809]: E1205 13:21:40.875542 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:21:53 crc kubenswrapper[4809]: I1205 13:21:53.872215 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:21:53 crc kubenswrapper[4809]: E1205 13:21:53.872894 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:22:04 crc kubenswrapper[4809]: I1205 13:22:04.886606 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:22:04 crc kubenswrapper[4809]: E1205 13:22:04.887278 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:22:15 crc kubenswrapper[4809]: I1205 13:22:15.874239 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:22:15 crc kubenswrapper[4809]: E1205 13:22:15.875399 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:22:30 crc kubenswrapper[4809]: I1205 13:22:30.877276 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:22:30 crc kubenswrapper[4809]: E1205 13:22:30.878376 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:22:39 crc kubenswrapper[4809]: I1205 13:22:39.824513 4809 generic.go:334] "Generic (PLEG): container finished" podID="328ce8d0-f42c-4e66-9de5-0f0cac6a2994" containerID="7548cc8b1a61d8e86f7e96ce1ef1e6b1970b18df61740eb1551ff22d1804ac8d" exitCode=0 Dec 05 13:22:39 crc kubenswrapper[4809]: I1205 13:22:39.824600 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" event={"ID":"328ce8d0-f42c-4e66-9de5-0f0cac6a2994","Type":"ContainerDied","Data":"7548cc8b1a61d8e86f7e96ce1ef1e6b1970b18df61740eb1551ff22d1804ac8d"} Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.334389 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.354480 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-secret-0\") pod \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.354626 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ceph\") pod \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.354731 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ssh-key\") pod \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.354867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-inventory\") pod \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.354926 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-combined-ca-bundle\") pod \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.354974 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4ztn\" (UniqueName: \"kubernetes.io/projected/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-kube-api-access-w4ztn\") pod \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\" (UID: \"328ce8d0-f42c-4e66-9de5-0f0cac6a2994\") " Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.360364 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "328ce8d0-f42c-4e66-9de5-0f0cac6a2994" (UID: "328ce8d0-f42c-4e66-9de5-0f0cac6a2994"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.366795 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ceph" (OuterVolumeSpecName: "ceph") pod "328ce8d0-f42c-4e66-9de5-0f0cac6a2994" (UID: "328ce8d0-f42c-4e66-9de5-0f0cac6a2994"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.376809 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-kube-api-access-w4ztn" (OuterVolumeSpecName: "kube-api-access-w4ztn") pod "328ce8d0-f42c-4e66-9de5-0f0cac6a2994" (UID: "328ce8d0-f42c-4e66-9de5-0f0cac6a2994"). InnerVolumeSpecName "kube-api-access-w4ztn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.397844 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "328ce8d0-f42c-4e66-9de5-0f0cac6a2994" (UID: "328ce8d0-f42c-4e66-9de5-0f0cac6a2994"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.408501 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-inventory" (OuterVolumeSpecName: "inventory") pod "328ce8d0-f42c-4e66-9de5-0f0cac6a2994" (UID: "328ce8d0-f42c-4e66-9de5-0f0cac6a2994"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.408954 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "328ce8d0-f42c-4e66-9de5-0f0cac6a2994" (UID: "328ce8d0-f42c-4e66-9de5-0f0cac6a2994"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.457715 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.457917 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.457992 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.458044 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.458113 4809 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.458168 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4ztn\" (UniqueName: \"kubernetes.io/projected/328ce8d0-f42c-4e66-9de5-0f0cac6a2994-kube-api-access-w4ztn\") on node \"crc\" DevicePath \"\"" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.855214 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" event={"ID":"328ce8d0-f42c-4e66-9de5-0f0cac6a2994","Type":"ContainerDied","Data":"c81ce1356c8952310f0bac91eac73e33597443783f58c5e52577762c1cc19f6a"} Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.855426 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c81ce1356c8952310f0bac91eac73e33597443783f58c5e52577762c1cc19f6a" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.855298 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-m8cjr" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.947961 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-gxdjl"] Dec 05 13:22:41 crc kubenswrapper[4809]: E1205 13:22:41.948472 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="extract-content" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.948491 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="extract-content" Dec 05 13:22:41 crc kubenswrapper[4809]: E1205 13:22:41.948505 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="registry-server" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.948520 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="registry-server" Dec 05 13:22:41 crc kubenswrapper[4809]: E1205 13:22:41.948528 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="extract-utilities" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.948536 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="extract-utilities" Dec 05 13:22:41 crc kubenswrapper[4809]: E1205 13:22:41.948551 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="328ce8d0-f42c-4e66-9de5-0f0cac6a2994" containerName="libvirt-openstack-openstack-cell1" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.948557 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="328ce8d0-f42c-4e66-9de5-0f0cac6a2994" containerName="libvirt-openstack-openstack-cell1" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.948805 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f4a69a-33ba-4652-8962-9179da9eb3d7" containerName="registry-server" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.948825 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="328ce8d0-f42c-4e66-9de5-0f0cac6a2994" containerName="libvirt-openstack-openstack-cell1" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.949555 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.953929 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.954024 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.954846 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.955291 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.955323 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.956163 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.956366 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:22:41 crc kubenswrapper[4809]: I1205 13:22:41.958286 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-gxdjl"] Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.002801 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.002973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ceph\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003107 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003145 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003297 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003369 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003434 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003552 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-inventory\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003590 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j8zw\" (UniqueName: \"kubernetes.io/projected/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-kube-api-access-9j8zw\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003615 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.003688 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105376 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105452 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105555 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105608 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ceph\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105668 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105689 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105732 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105841 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-inventory\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.105861 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j8zw\" (UniqueName: \"kubernetes.io/projected/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-kube-api-access-9j8zw\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.106974 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.109117 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.110562 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.111566 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ceph\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.112143 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.115138 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.115203 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.116192 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.119102 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.123385 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-inventory\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.129603 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j8zw\" (UniqueName: \"kubernetes.io/projected/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-kube-api-access-9j8zw\") pod \"nova-cell1-openstack-openstack-cell1-gxdjl\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.269412 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.848701 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-gxdjl"] Dec 05 13:22:42 crc kubenswrapper[4809]: I1205 13:22:42.870840 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" event={"ID":"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2","Type":"ContainerStarted","Data":"5a5b80133775dc41de47b5aba19ac5fe9098d54b4604eb294916caa75393126d"} Dec 05 13:22:43 crc kubenswrapper[4809]: I1205 13:22:43.881514 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" event={"ID":"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2","Type":"ContainerStarted","Data":"7892c0b58783da22d717435719b1eb72950665907dafa1d665ef6d4bb7764379"} Dec 05 13:22:45 crc kubenswrapper[4809]: I1205 13:22:45.874953 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:22:45 crc kubenswrapper[4809]: E1205 13:22:45.875484 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:23:00 crc kubenswrapper[4809]: I1205 13:23:00.873050 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:23:00 crc kubenswrapper[4809]: E1205 13:23:00.873945 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:23:14 crc kubenswrapper[4809]: I1205 13:23:14.878052 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:23:15 crc kubenswrapper[4809]: I1205 13:23:15.219872 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"9d3685e851dddaeba4146f02e33020bb81f8608a4e56185eb1fe00af2610068a"} Dec 05 13:23:15 crc kubenswrapper[4809]: I1205 13:23:15.247779 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" podStartSLOduration=33.657696293 podStartE2EDuration="34.247761465s" podCreationTimestamp="2025-12-05 13:22:41 +0000 UTC" firstStartedPulling="2025-12-05 13:22:42.853985373 +0000 UTC m=+8058.244961941" lastFinishedPulling="2025-12-05 13:22:43.444050555 +0000 UTC m=+8058.835027113" observedRunningTime="2025-12-05 13:22:43.906123785 +0000 UTC m=+8059.297100363" watchObservedRunningTime="2025-12-05 13:23:15.247761465 +0000 UTC m=+8090.638738013" Dec 05 13:23:29 crc kubenswrapper[4809]: I1205 13:23:29.815688 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tjcng"] Dec 05 13:23:29 crc kubenswrapper[4809]: I1205 13:23:29.820649 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:29 crc kubenswrapper[4809]: I1205 13:23:29.831734 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tjcng"] Dec 05 13:23:29 crc kubenswrapper[4809]: I1205 13:23:29.917035 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-catalog-content\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:29 crc kubenswrapper[4809]: I1205 13:23:29.917095 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hg4p\" (UniqueName: \"kubernetes.io/projected/465d79fc-4f66-4371-b2fb-c195d4c590fe-kube-api-access-9hg4p\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:29 crc kubenswrapper[4809]: I1205 13:23:29.917146 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-utilities\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.019394 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-catalog-content\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.019797 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hg4p\" (UniqueName: \"kubernetes.io/projected/465d79fc-4f66-4371-b2fb-c195d4c590fe-kube-api-access-9hg4p\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.019946 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-utilities\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.020495 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-catalog-content\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.020660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-utilities\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.042620 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hg4p\" (UniqueName: \"kubernetes.io/projected/465d79fc-4f66-4371-b2fb-c195d4c590fe-kube-api-access-9hg4p\") pod \"redhat-marketplace-tjcng\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.185205 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:30 crc kubenswrapper[4809]: I1205 13:23:30.780917 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tjcng"] Dec 05 13:23:31 crc kubenswrapper[4809]: I1205 13:23:31.405601 4809 generic.go:334] "Generic (PLEG): container finished" podID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerID="5177391db0a34b697e4ed364098ecf4ee8c876ebd6b9c515ecec69d295b921ea" exitCode=0 Dec 05 13:23:31 crc kubenswrapper[4809]: I1205 13:23:31.405701 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjcng" event={"ID":"465d79fc-4f66-4371-b2fb-c195d4c590fe","Type":"ContainerDied","Data":"5177391db0a34b697e4ed364098ecf4ee8c876ebd6b9c515ecec69d295b921ea"} Dec 05 13:23:31 crc kubenswrapper[4809]: I1205 13:23:31.405738 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjcng" event={"ID":"465d79fc-4f66-4371-b2fb-c195d4c590fe","Type":"ContainerStarted","Data":"e6bbdba242b94fc2023f15b3c668b4287fa7bbc1c6ba31fd075bf9374b6c8035"} Dec 05 13:23:32 crc kubenswrapper[4809]: I1205 13:23:32.419933 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjcng" event={"ID":"465d79fc-4f66-4371-b2fb-c195d4c590fe","Type":"ContainerStarted","Data":"931a417e7468b569645edcdebb0cd5e56c66395d6cdc68219927ca4ed2bff88a"} Dec 05 13:23:33 crc kubenswrapper[4809]: I1205 13:23:33.436274 4809 generic.go:334] "Generic (PLEG): container finished" podID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerID="931a417e7468b569645edcdebb0cd5e56c66395d6cdc68219927ca4ed2bff88a" exitCode=0 Dec 05 13:23:33 crc kubenswrapper[4809]: I1205 13:23:33.436921 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjcng" event={"ID":"465d79fc-4f66-4371-b2fb-c195d4c590fe","Type":"ContainerDied","Data":"931a417e7468b569645edcdebb0cd5e56c66395d6cdc68219927ca4ed2bff88a"} Dec 05 13:23:34 crc kubenswrapper[4809]: I1205 13:23:34.448076 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjcng" event={"ID":"465d79fc-4f66-4371-b2fb-c195d4c590fe","Type":"ContainerStarted","Data":"48bfcf663cccf04c94ea2c24bbd10dc0322f667a0b3cfa763f11d7b166974925"} Dec 05 13:23:34 crc kubenswrapper[4809]: I1205 13:23:34.469994 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tjcng" podStartSLOduration=2.984857732 podStartE2EDuration="5.469977871s" podCreationTimestamp="2025-12-05 13:23:29 +0000 UTC" firstStartedPulling="2025-12-05 13:23:31.407998335 +0000 UTC m=+8106.798974903" lastFinishedPulling="2025-12-05 13:23:33.893118464 +0000 UTC m=+8109.284095042" observedRunningTime="2025-12-05 13:23:34.46623689 +0000 UTC m=+8109.857213458" watchObservedRunningTime="2025-12-05 13:23:34.469977871 +0000 UTC m=+8109.860954429" Dec 05 13:23:40 crc kubenswrapper[4809]: I1205 13:23:40.186050 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:40 crc kubenswrapper[4809]: I1205 13:23:40.186533 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:40 crc kubenswrapper[4809]: I1205 13:23:40.251973 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:40 crc kubenswrapper[4809]: I1205 13:23:40.600472 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:40 crc kubenswrapper[4809]: I1205 13:23:40.661134 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tjcng"] Dec 05 13:23:42 crc kubenswrapper[4809]: I1205 13:23:42.550981 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tjcng" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="registry-server" containerID="cri-o://48bfcf663cccf04c94ea2c24bbd10dc0322f667a0b3cfa763f11d7b166974925" gracePeriod=2 Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.567596 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjcng" event={"ID":"465d79fc-4f66-4371-b2fb-c195d4c590fe","Type":"ContainerDied","Data":"48bfcf663cccf04c94ea2c24bbd10dc0322f667a0b3cfa763f11d7b166974925"} Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.567512 4809 generic.go:334] "Generic (PLEG): container finished" podID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerID="48bfcf663cccf04c94ea2c24bbd10dc0322f667a0b3cfa763f11d7b166974925" exitCode=0 Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.666965 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.739849 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-catalog-content\") pod \"465d79fc-4f66-4371-b2fb-c195d4c590fe\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.740248 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hg4p\" (UniqueName: \"kubernetes.io/projected/465d79fc-4f66-4371-b2fb-c195d4c590fe-kube-api-access-9hg4p\") pod \"465d79fc-4f66-4371-b2fb-c195d4c590fe\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.740303 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-utilities\") pod \"465d79fc-4f66-4371-b2fb-c195d4c590fe\" (UID: \"465d79fc-4f66-4371-b2fb-c195d4c590fe\") " Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.748945 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/465d79fc-4f66-4371-b2fb-c195d4c590fe-kube-api-access-9hg4p" (OuterVolumeSpecName: "kube-api-access-9hg4p") pod "465d79fc-4f66-4371-b2fb-c195d4c590fe" (UID: "465d79fc-4f66-4371-b2fb-c195d4c590fe"). InnerVolumeSpecName "kube-api-access-9hg4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.753099 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-utilities" (OuterVolumeSpecName: "utilities") pod "465d79fc-4f66-4371-b2fb-c195d4c590fe" (UID: "465d79fc-4f66-4371-b2fb-c195d4c590fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.769321 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "465d79fc-4f66-4371-b2fb-c195d4c590fe" (UID: "465d79fc-4f66-4371-b2fb-c195d4c590fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.843121 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.843163 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/465d79fc-4f66-4371-b2fb-c195d4c590fe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:23:43 crc kubenswrapper[4809]: I1205 13:23:43.843180 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hg4p\" (UniqueName: \"kubernetes.io/projected/465d79fc-4f66-4371-b2fb-c195d4c590fe-kube-api-access-9hg4p\") on node \"crc\" DevicePath \"\"" Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.583544 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tjcng" event={"ID":"465d79fc-4f66-4371-b2fb-c195d4c590fe","Type":"ContainerDied","Data":"e6bbdba242b94fc2023f15b3c668b4287fa7bbc1c6ba31fd075bf9374b6c8035"} Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.583617 4809 scope.go:117] "RemoveContainer" containerID="48bfcf663cccf04c94ea2c24bbd10dc0322f667a0b3cfa763f11d7b166974925" Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.583683 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tjcng" Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.635529 4809 scope.go:117] "RemoveContainer" containerID="931a417e7468b569645edcdebb0cd5e56c66395d6cdc68219927ca4ed2bff88a" Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.670454 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tjcng"] Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.710785 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tjcng"] Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.710946 4809 scope.go:117] "RemoveContainer" containerID="5177391db0a34b697e4ed364098ecf4ee8c876ebd6b9c515ecec69d295b921ea" Dec 05 13:23:44 crc kubenswrapper[4809]: I1205 13:23:44.885728 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" path="/var/lib/kubelet/pods/465d79fc-4f66-4371-b2fb-c195d4c590fe/volumes" Dec 05 13:24:45 crc kubenswrapper[4809]: I1205 13:24:45.524879 4809 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-7n7zc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 13:24:45 crc kubenswrapper[4809]: I1205 13:24:45.525408 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7n7zc" podUID="ddd951b5-5bbf-4f65-b09a-4482bb58fcb6" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 13:25:44 crc kubenswrapper[4809]: I1205 13:25:44.046378 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:25:44 crc kubenswrapper[4809]: I1205 13:25:44.047267 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:25:56 crc kubenswrapper[4809]: I1205 13:25:56.637807 4809 generic.go:334] "Generic (PLEG): container finished" podID="d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" containerID="7892c0b58783da22d717435719b1eb72950665907dafa1d665ef6d4bb7764379" exitCode=0 Dec 05 13:25:56 crc kubenswrapper[4809]: I1205 13:25:56.637895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" event={"ID":"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2","Type":"ContainerDied","Data":"7892c0b58783da22d717435719b1eb72950665907dafa1d665ef6d4bb7764379"} Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.133466 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.309704 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ceph\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.309896 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-0\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.309950 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-combined-ca-bundle\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.309985 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-inventory\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.310059 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-1\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.310080 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-0\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.310099 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ssh-key\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.310145 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-1\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.310161 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-1\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.310194 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j8zw\" (UniqueName: \"kubernetes.io/projected/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-kube-api-access-9j8zw\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.310213 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-0\") pod \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\" (UID: \"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2\") " Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.317525 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-kube-api-access-9j8zw" (OuterVolumeSpecName: "kube-api-access-9j8zw") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "kube-api-access-9j8zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.333053 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ceph" (OuterVolumeSpecName: "ceph") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.333187 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.349123 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.357271 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.358782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.361034 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.361389 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.363123 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-inventory" (OuterVolumeSpecName: "inventory") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.366435 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.392307 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" (UID: "d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412299 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412336 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412348 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412359 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412370 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412381 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412390 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412407 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412415 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j8zw\" (UniqueName: \"kubernetes.io/projected/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-kube-api-access-9j8zw\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412424 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.412432 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.658767 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" event={"ID":"d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2","Type":"ContainerDied","Data":"5a5b80133775dc41de47b5aba19ac5fe9098d54b4604eb294916caa75393126d"} Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.659075 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a5b80133775dc41de47b5aba19ac5fe9098d54b4604eb294916caa75393126d" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.658864 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-gxdjl" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.789179 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-7mb5p"] Dec 05 13:25:58 crc kubenswrapper[4809]: E1205 13:25:58.789773 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="extract-utilities" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.789794 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="extract-utilities" Dec 05 13:25:58 crc kubenswrapper[4809]: E1205 13:25:58.789823 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="extract-content" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.789834 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="extract-content" Dec 05 13:25:58 crc kubenswrapper[4809]: E1205 13:25:58.789863 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="registry-server" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.789872 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="registry-server" Dec 05 13:25:58 crc kubenswrapper[4809]: E1205 13:25:58.789884 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" containerName="nova-cell1-openstack-openstack-cell1" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.789892 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" containerName="nova-cell1-openstack-openstack-cell1" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.790155 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="465d79fc-4f66-4371-b2fb-c195d4c590fe" containerName="registry-server" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.790177 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2" containerName="nova-cell1-openstack-openstack-cell1" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.791222 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.794591 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.794759 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.794836 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.795508 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.797025 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.803374 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-7mb5p"] Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949546 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgl7g\" (UniqueName: \"kubernetes.io/projected/0fb44e23-96d1-4a58-b7df-0522db74bb22-kube-api-access-tgl7g\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949673 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceph\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949734 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949782 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ssh-key\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949898 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949943 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-inventory\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:58 crc kubenswrapper[4809]: I1205 13:25:58.949985 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.052786 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceph\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.053896 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.054000 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.054293 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ssh-key\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.054399 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.054499 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-inventory\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.054587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.055720 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgl7g\" (UniqueName: \"kubernetes.io/projected/0fb44e23-96d1-4a58-b7df-0522db74bb22-kube-api-access-tgl7g\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.057330 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.058312 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceph\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.060311 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.061588 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.063762 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.073305 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-inventory\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.076966 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ssh-key\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.078415 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgl7g\" (UniqueName: \"kubernetes.io/projected/0fb44e23-96d1-4a58-b7df-0522db74bb22-kube-api-access-tgl7g\") pod \"telemetry-openstack-openstack-cell1-7mb5p\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.177646 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:25:59 crc kubenswrapper[4809]: I1205 13:25:59.796250 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-7mb5p"] Dec 05 13:26:00 crc kubenswrapper[4809]: I1205 13:26:00.679092 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" event={"ID":"0fb44e23-96d1-4a58-b7df-0522db74bb22","Type":"ContainerStarted","Data":"341e14689a335ff8527d0aa878a2b640ef9aa9bcb0da603e5bb099ed53c5e656"} Dec 05 13:26:03 crc kubenswrapper[4809]: I1205 13:26:03.805275 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" event={"ID":"0fb44e23-96d1-4a58-b7df-0522db74bb22","Type":"ContainerStarted","Data":"5a8f598ffdb56508ad8725a6cf2bede703a2d436f46b7464fb64b391b4238a86"} Dec 05 13:26:03 crc kubenswrapper[4809]: I1205 13:26:03.828520 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" podStartSLOduration=3.08826094 podStartE2EDuration="5.82849218s" podCreationTimestamp="2025-12-05 13:25:58 +0000 UTC" firstStartedPulling="2025-12-05 13:25:59.794157247 +0000 UTC m=+8255.185133795" lastFinishedPulling="2025-12-05 13:26:02.534388457 +0000 UTC m=+8257.925365035" observedRunningTime="2025-12-05 13:26:03.826882957 +0000 UTC m=+8259.217859585" watchObservedRunningTime="2025-12-05 13:26:03.82849218 +0000 UTC m=+8259.219468778" Dec 05 13:26:14 crc kubenswrapper[4809]: I1205 13:26:14.046609 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:26:14 crc kubenswrapper[4809]: I1205 13:26:14.047852 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.046460 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.047282 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.047359 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.048291 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9d3685e851dddaeba4146f02e33020bb81f8608a4e56185eb1fe00af2610068a"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.048360 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://9d3685e851dddaeba4146f02e33020bb81f8608a4e56185eb1fe00af2610068a" gracePeriod=600 Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.291080 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="9d3685e851dddaeba4146f02e33020bb81f8608a4e56185eb1fe00af2610068a" exitCode=0 Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.291157 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"9d3685e851dddaeba4146f02e33020bb81f8608a4e56185eb1fe00af2610068a"} Dec 05 13:26:44 crc kubenswrapper[4809]: I1205 13:26:44.291210 4809 scope.go:117] "RemoveContainer" containerID="1549213cc4e78897642f60b5cd444ec4d8d0ed290def20733777cbd4637c2c1f" Dec 05 13:26:45 crc kubenswrapper[4809]: I1205 13:26:45.309941 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c"} Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.405580 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g2sdw"] Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.409982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.423859 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g2sdw"] Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.563465 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-catalog-content\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.563870 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-utilities\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.564158 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfbt5\" (UniqueName: \"kubernetes.io/projected/80d637ae-2388-4a6b-829d-724eb50c26f3-kube-api-access-zfbt5\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.666015 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-catalog-content\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.666111 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-utilities\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.666215 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfbt5\" (UniqueName: \"kubernetes.io/projected/80d637ae-2388-4a6b-829d-724eb50c26f3-kube-api-access-zfbt5\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.666605 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-catalog-content\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.666806 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-utilities\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.689707 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfbt5\" (UniqueName: \"kubernetes.io/projected/80d637ae-2388-4a6b-829d-724eb50c26f3-kube-api-access-zfbt5\") pod \"redhat-operators-g2sdw\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:47 crc kubenswrapper[4809]: I1205 13:26:47.742493 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:26:48 crc kubenswrapper[4809]: I1205 13:26:48.282058 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g2sdw"] Dec 05 13:26:48 crc kubenswrapper[4809]: I1205 13:26:48.355499 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2sdw" event={"ID":"80d637ae-2388-4a6b-829d-724eb50c26f3","Type":"ContainerStarted","Data":"66acaf1eb87de33b3e8a72e058159633047d855da74592556f8ee7d61e23d341"} Dec 05 13:26:49 crc kubenswrapper[4809]: I1205 13:26:49.373678 4809 generic.go:334] "Generic (PLEG): container finished" podID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerID="b1046a11182416ba2eb66545283568660f1211b92696877c54a4da4196ef5051" exitCode=0 Dec 05 13:26:49 crc kubenswrapper[4809]: I1205 13:26:49.373772 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2sdw" event={"ID":"80d637ae-2388-4a6b-829d-724eb50c26f3","Type":"ContainerDied","Data":"b1046a11182416ba2eb66545283568660f1211b92696877c54a4da4196ef5051"} Dec 05 13:26:49 crc kubenswrapper[4809]: I1205 13:26:49.386350 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:26:51 crc kubenswrapper[4809]: I1205 13:26:51.398463 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2sdw" event={"ID":"80d637ae-2388-4a6b-829d-724eb50c26f3","Type":"ContainerStarted","Data":"cdc36781f520d2eef6fbe8e58c93de3f71ee765d3fa2662fe18cd1c787c6d10c"} Dec 05 13:27:01 crc kubenswrapper[4809]: I1205 13:27:01.520850 4809 generic.go:334] "Generic (PLEG): container finished" podID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerID="cdc36781f520d2eef6fbe8e58c93de3f71ee765d3fa2662fe18cd1c787c6d10c" exitCode=0 Dec 05 13:27:01 crc kubenswrapper[4809]: I1205 13:27:01.520912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2sdw" event={"ID":"80d637ae-2388-4a6b-829d-724eb50c26f3","Type":"ContainerDied","Data":"cdc36781f520d2eef6fbe8e58c93de3f71ee765d3fa2662fe18cd1c787c6d10c"} Dec 05 13:27:03 crc kubenswrapper[4809]: I1205 13:27:03.550795 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2sdw" event={"ID":"80d637ae-2388-4a6b-829d-724eb50c26f3","Type":"ContainerStarted","Data":"cc55786aa1907d006a563fa53d2a700313ae8d191061646ab998b605f0c40593"} Dec 05 13:27:03 crc kubenswrapper[4809]: I1205 13:27:03.584395 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g2sdw" podStartSLOduration=3.296798502 podStartE2EDuration="16.58436841s" podCreationTimestamp="2025-12-05 13:26:47 +0000 UTC" firstStartedPulling="2025-12-05 13:26:49.385967533 +0000 UTC m=+8304.776944101" lastFinishedPulling="2025-12-05 13:27:02.673537451 +0000 UTC m=+8318.064514009" observedRunningTime="2025-12-05 13:27:03.577819606 +0000 UTC m=+8318.968796224" watchObservedRunningTime="2025-12-05 13:27:03.58436841 +0000 UTC m=+8318.975344988" Dec 05 13:27:07 crc kubenswrapper[4809]: I1205 13:27:07.743528 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:27:07 crc kubenswrapper[4809]: I1205 13:27:07.744148 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:27:08 crc kubenswrapper[4809]: I1205 13:27:08.832106 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g2sdw" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="registry-server" probeResult="failure" output=< Dec 05 13:27:08 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 13:27:08 crc kubenswrapper[4809]: > Dec 05 13:27:17 crc kubenswrapper[4809]: I1205 13:27:17.807879 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:27:17 crc kubenswrapper[4809]: I1205 13:27:17.873057 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:27:18 crc kubenswrapper[4809]: I1205 13:27:18.599829 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g2sdw"] Dec 05 13:27:19 crc kubenswrapper[4809]: I1205 13:27:19.748622 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g2sdw" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="registry-server" containerID="cri-o://cc55786aa1907d006a563fa53d2a700313ae8d191061646ab998b605f0c40593" gracePeriod=2 Dec 05 13:27:20 crc kubenswrapper[4809]: I1205 13:27:20.762557 4809 generic.go:334] "Generic (PLEG): container finished" podID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerID="cc55786aa1907d006a563fa53d2a700313ae8d191061646ab998b605f0c40593" exitCode=0 Dec 05 13:27:20 crc kubenswrapper[4809]: I1205 13:27:20.762762 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2sdw" event={"ID":"80d637ae-2388-4a6b-829d-724eb50c26f3","Type":"ContainerDied","Data":"cc55786aa1907d006a563fa53d2a700313ae8d191061646ab998b605f0c40593"} Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.526184 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.565854 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-catalog-content\") pod \"80d637ae-2388-4a6b-829d-724eb50c26f3\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.566201 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfbt5\" (UniqueName: \"kubernetes.io/projected/80d637ae-2388-4a6b-829d-724eb50c26f3-kube-api-access-zfbt5\") pod \"80d637ae-2388-4a6b-829d-724eb50c26f3\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.566328 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-utilities\") pod \"80d637ae-2388-4a6b-829d-724eb50c26f3\" (UID: \"80d637ae-2388-4a6b-829d-724eb50c26f3\") " Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.567138 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-utilities" (OuterVolumeSpecName: "utilities") pod "80d637ae-2388-4a6b-829d-724eb50c26f3" (UID: "80d637ae-2388-4a6b-829d-724eb50c26f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.573923 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80d637ae-2388-4a6b-829d-724eb50c26f3-kube-api-access-zfbt5" (OuterVolumeSpecName: "kube-api-access-zfbt5") pod "80d637ae-2388-4a6b-829d-724eb50c26f3" (UID: "80d637ae-2388-4a6b-829d-724eb50c26f3"). InnerVolumeSpecName "kube-api-access-zfbt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.574768 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfbt5\" (UniqueName: \"kubernetes.io/projected/80d637ae-2388-4a6b-829d-724eb50c26f3-kube-api-access-zfbt5\") on node \"crc\" DevicePath \"\"" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.574792 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.694583 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80d637ae-2388-4a6b-829d-724eb50c26f3" (UID: "80d637ae-2388-4a6b-829d-724eb50c26f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.777346 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2sdw" event={"ID":"80d637ae-2388-4a6b-829d-724eb50c26f3","Type":"ContainerDied","Data":"66acaf1eb87de33b3e8a72e058159633047d855da74592556f8ee7d61e23d341"} Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.777400 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2sdw" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.777419 4809 scope.go:117] "RemoveContainer" containerID="cc55786aa1907d006a563fa53d2a700313ae8d191061646ab998b605f0c40593" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.782355 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d637ae-2388-4a6b-829d-724eb50c26f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.810399 4809 scope.go:117] "RemoveContainer" containerID="cdc36781f520d2eef6fbe8e58c93de3f71ee765d3fa2662fe18cd1c787c6d10c" Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.818713 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g2sdw"] Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.830463 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g2sdw"] Dec 05 13:27:21 crc kubenswrapper[4809]: I1205 13:27:21.849853 4809 scope.go:117] "RemoveContainer" containerID="b1046a11182416ba2eb66545283568660f1211b92696877c54a4da4196ef5051" Dec 05 13:27:22 crc kubenswrapper[4809]: I1205 13:27:22.884281 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" path="/var/lib/kubelet/pods/80d637ae-2388-4a6b-829d-724eb50c26f3/volumes" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.642728 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qgdp6"] Dec 05 13:28:26 crc kubenswrapper[4809]: E1205 13:28:26.643731 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="extract-content" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.643744 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="extract-content" Dec 05 13:28:26 crc kubenswrapper[4809]: E1205 13:28:26.643759 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="registry-server" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.643769 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="registry-server" Dec 05 13:28:26 crc kubenswrapper[4809]: E1205 13:28:26.643806 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="extract-utilities" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.643813 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="extract-utilities" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.644038 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="80d637ae-2388-4a6b-829d-724eb50c26f3" containerName="registry-server" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.645946 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.714545 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgdp6"] Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.723767 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9qzz\" (UniqueName: \"kubernetes.io/projected/ae8dd80c-4350-488a-a4ff-010ed61309d1-kube-api-access-g9qzz\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.723904 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-utilities\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.724068 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-catalog-content\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.829541 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-catalog-content\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.829617 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9qzz\" (UniqueName: \"kubernetes.io/projected/ae8dd80c-4350-488a-a4ff-010ed61309d1-kube-api-access-g9qzz\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.829702 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-utilities\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.830177 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-utilities\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.830308 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-catalog-content\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.859876 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9qzz\" (UniqueName: \"kubernetes.io/projected/ae8dd80c-4350-488a-a4ff-010ed61309d1-kube-api-access-g9qzz\") pod \"community-operators-qgdp6\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:26 crc kubenswrapper[4809]: I1205 13:28:26.980701 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:27 crc kubenswrapper[4809]: I1205 13:28:27.522797 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgdp6"] Dec 05 13:28:27 crc kubenswrapper[4809]: I1205 13:28:27.590239 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgdp6" event={"ID":"ae8dd80c-4350-488a-a4ff-010ed61309d1","Type":"ContainerStarted","Data":"1e17ce0f3caabf9e224cfc2131aa60bbd426af11e63e23dc481869c2ff5b6fa6"} Dec 05 13:28:28 crc kubenswrapper[4809]: I1205 13:28:28.601678 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerID="e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63" exitCode=0 Dec 05 13:28:28 crc kubenswrapper[4809]: I1205 13:28:28.601719 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgdp6" event={"ID":"ae8dd80c-4350-488a-a4ff-010ed61309d1","Type":"ContainerDied","Data":"e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63"} Dec 05 13:28:30 crc kubenswrapper[4809]: I1205 13:28:30.625976 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerID="7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f" exitCode=0 Dec 05 13:28:30 crc kubenswrapper[4809]: I1205 13:28:30.626033 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgdp6" event={"ID":"ae8dd80c-4350-488a-a4ff-010ed61309d1","Type":"ContainerDied","Data":"7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f"} Dec 05 13:28:31 crc kubenswrapper[4809]: I1205 13:28:31.640678 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgdp6" event={"ID":"ae8dd80c-4350-488a-a4ff-010ed61309d1","Type":"ContainerStarted","Data":"5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c"} Dec 05 13:28:31 crc kubenswrapper[4809]: I1205 13:28:31.664350 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qgdp6" podStartSLOduration=2.94385984 podStartE2EDuration="5.664327578s" podCreationTimestamp="2025-12-05 13:28:26 +0000 UTC" firstStartedPulling="2025-12-05 13:28:28.605033574 +0000 UTC m=+8403.996010172" lastFinishedPulling="2025-12-05 13:28:31.325501342 +0000 UTC m=+8406.716477910" observedRunningTime="2025-12-05 13:28:31.660018394 +0000 UTC m=+8407.050994972" watchObservedRunningTime="2025-12-05 13:28:31.664327578 +0000 UTC m=+8407.055304136" Dec 05 13:28:36 crc kubenswrapper[4809]: I1205 13:28:36.981462 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:36 crc kubenswrapper[4809]: I1205 13:28:36.982351 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:37 crc kubenswrapper[4809]: I1205 13:28:37.044863 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:37 crc kubenswrapper[4809]: I1205 13:28:37.789491 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:37 crc kubenswrapper[4809]: I1205 13:28:37.850808 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgdp6"] Dec 05 13:28:39 crc kubenswrapper[4809]: I1205 13:28:39.734278 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qgdp6" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="registry-server" containerID="cri-o://5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c" gracePeriod=2 Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.292558 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.352885 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-catalog-content\") pod \"ae8dd80c-4350-488a-a4ff-010ed61309d1\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.353010 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-utilities\") pod \"ae8dd80c-4350-488a-a4ff-010ed61309d1\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.353194 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9qzz\" (UniqueName: \"kubernetes.io/projected/ae8dd80c-4350-488a-a4ff-010ed61309d1-kube-api-access-g9qzz\") pod \"ae8dd80c-4350-488a-a4ff-010ed61309d1\" (UID: \"ae8dd80c-4350-488a-a4ff-010ed61309d1\") " Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.354278 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-utilities" (OuterVolumeSpecName: "utilities") pod "ae8dd80c-4350-488a-a4ff-010ed61309d1" (UID: "ae8dd80c-4350-488a-a4ff-010ed61309d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.360744 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8dd80c-4350-488a-a4ff-010ed61309d1-kube-api-access-g9qzz" (OuterVolumeSpecName: "kube-api-access-g9qzz") pod "ae8dd80c-4350-488a-a4ff-010ed61309d1" (UID: "ae8dd80c-4350-488a-a4ff-010ed61309d1"). InnerVolumeSpecName "kube-api-access-g9qzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.412906 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae8dd80c-4350-488a-a4ff-010ed61309d1" (UID: "ae8dd80c-4350-488a-a4ff-010ed61309d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.455911 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.455965 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8dd80c-4350-488a-a4ff-010ed61309d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.455983 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9qzz\" (UniqueName: \"kubernetes.io/projected/ae8dd80c-4350-488a-a4ff-010ed61309d1-kube-api-access-g9qzz\") on node \"crc\" DevicePath \"\"" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.748041 4809 generic.go:334] "Generic (PLEG): container finished" podID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerID="5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c" exitCode=0 Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.748112 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgdp6" event={"ID":"ae8dd80c-4350-488a-a4ff-010ed61309d1","Type":"ContainerDied","Data":"5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c"} Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.748145 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgdp6" event={"ID":"ae8dd80c-4350-488a-a4ff-010ed61309d1","Type":"ContainerDied","Data":"1e17ce0f3caabf9e224cfc2131aa60bbd426af11e63e23dc481869c2ff5b6fa6"} Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.748165 4809 scope.go:117] "RemoveContainer" containerID="5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.748164 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgdp6" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.779809 4809 scope.go:117] "RemoveContainer" containerID="7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.812077 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgdp6"] Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.817566 4809 scope.go:117] "RemoveContainer" containerID="e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.825996 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qgdp6"] Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.901690 4809 scope.go:117] "RemoveContainer" containerID="5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c" Dec 05 13:28:40 crc kubenswrapper[4809]: E1205 13:28:40.902764 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c\": container with ID starting with 5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c not found: ID does not exist" containerID="5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.902811 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c"} err="failed to get container status \"5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c\": rpc error: code = NotFound desc = could not find container \"5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c\": container with ID starting with 5ff4638e9437e2996e71ea9aa92440c0d1469dd7b27d8525da87ce4007cde89c not found: ID does not exist" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.902848 4809 scope.go:117] "RemoveContainer" containerID="7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f" Dec 05 13:28:40 crc kubenswrapper[4809]: E1205 13:28:40.903179 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f\": container with ID starting with 7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f not found: ID does not exist" containerID="7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.903218 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f"} err="failed to get container status \"7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f\": rpc error: code = NotFound desc = could not find container \"7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f\": container with ID starting with 7587645b27f3f106366cb3c18621f6d840d2c8ef08d7e4e29f9c41c894272b5f not found: ID does not exist" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.903240 4809 scope.go:117] "RemoveContainer" containerID="e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63" Dec 05 13:28:40 crc kubenswrapper[4809]: E1205 13:28:40.903526 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63\": container with ID starting with e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63 not found: ID does not exist" containerID="e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.903563 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63"} err="failed to get container status \"e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63\": rpc error: code = NotFound desc = could not find container \"e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63\": container with ID starting with e5097a28d9a54727f138be7996987951138c976fa25b298e2feb47c74159da63 not found: ID does not exist" Dec 05 13:28:40 crc kubenswrapper[4809]: I1205 13:28:40.920526 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" path="/var/lib/kubelet/pods/ae8dd80c-4350-488a-a4ff-010ed61309d1/volumes" Dec 05 13:28:44 crc kubenswrapper[4809]: I1205 13:28:44.046408 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:28:44 crc kubenswrapper[4809]: I1205 13:28:44.046844 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:29:14 crc kubenswrapper[4809]: I1205 13:29:14.046778 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:29:14 crc kubenswrapper[4809]: I1205 13:29:14.047387 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.047876 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.048940 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.049006 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.050106 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.050189 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" gracePeriod=600 Dec 05 13:29:44 crc kubenswrapper[4809]: E1205 13:29:44.181902 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.434061 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" exitCode=0 Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.434120 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c"} Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.434160 4809 scope.go:117] "RemoveContainer" containerID="9d3685e851dddaeba4146f02e33020bb81f8608a4e56185eb1fe00af2610068a" Dec 05 13:29:44 crc kubenswrapper[4809]: I1205 13:29:44.434872 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:29:44 crc kubenswrapper[4809]: E1205 13:29:44.435175 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:29:59 crc kubenswrapper[4809]: I1205 13:29:59.872995 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:29:59 crc kubenswrapper[4809]: E1205 13:29:59.873870 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.164424 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7"] Dec 05 13:30:00 crc kubenswrapper[4809]: E1205 13:30:00.165300 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="registry-server" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.165325 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="registry-server" Dec 05 13:30:00 crc kubenswrapper[4809]: E1205 13:30:00.165345 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="extract-content" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.165354 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="extract-content" Dec 05 13:30:00 crc kubenswrapper[4809]: E1205 13:30:00.165417 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="extract-utilities" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.165428 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="extract-utilities" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.165695 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8dd80c-4350-488a-a4ff-010ed61309d1" containerName="registry-server" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.166768 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.171089 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.171257 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.181836 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7"] Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.269973 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273ff4c8-4957-434f-a9f6-26ef0ed0f478-config-volume\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.270172 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9fwn\" (UniqueName: \"kubernetes.io/projected/273ff4c8-4957-434f-a9f6-26ef0ed0f478-kube-api-access-q9fwn\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.270347 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273ff4c8-4957-434f-a9f6-26ef0ed0f478-secret-volume\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.372426 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273ff4c8-4957-434f-a9f6-26ef0ed0f478-secret-volume\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.372579 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273ff4c8-4957-434f-a9f6-26ef0ed0f478-config-volume\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.372682 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9fwn\" (UniqueName: \"kubernetes.io/projected/273ff4c8-4957-434f-a9f6-26ef0ed0f478-kube-api-access-q9fwn\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.373570 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273ff4c8-4957-434f-a9f6-26ef0ed0f478-config-volume\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.380084 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273ff4c8-4957-434f-a9f6-26ef0ed0f478-secret-volume\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.391931 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9fwn\" (UniqueName: \"kubernetes.io/projected/273ff4c8-4957-434f-a9f6-26ef0ed0f478-kube-api-access-q9fwn\") pod \"collect-profiles-29415690-wx4h7\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:00 crc kubenswrapper[4809]: I1205 13:30:00.495375 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:01 crc kubenswrapper[4809]: I1205 13:30:01.009945 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7"] Dec 05 13:30:01 crc kubenswrapper[4809]: I1205 13:30:01.658314 4809 generic.go:334] "Generic (PLEG): container finished" podID="273ff4c8-4957-434f-a9f6-26ef0ed0f478" containerID="ce2cfc41ac01a572fa1bdc6328610a1f7d8c8ebc23981aac37e9f0b820c5f052" exitCode=0 Dec 05 13:30:01 crc kubenswrapper[4809]: I1205 13:30:01.658391 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" event={"ID":"273ff4c8-4957-434f-a9f6-26ef0ed0f478","Type":"ContainerDied","Data":"ce2cfc41ac01a572fa1bdc6328610a1f7d8c8ebc23981aac37e9f0b820c5f052"} Dec 05 13:30:01 crc kubenswrapper[4809]: I1205 13:30:01.658591 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" event={"ID":"273ff4c8-4957-434f-a9f6-26ef0ed0f478","Type":"ContainerStarted","Data":"95015be276e81c1a382def47c5869241cf8dee38c03886e51e2981791c45edf1"} Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.069828 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.231856 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273ff4c8-4957-434f-a9f6-26ef0ed0f478-secret-volume\") pod \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.232011 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273ff4c8-4957-434f-a9f6-26ef0ed0f478-config-volume\") pod \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.232274 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9fwn\" (UniqueName: \"kubernetes.io/projected/273ff4c8-4957-434f-a9f6-26ef0ed0f478-kube-api-access-q9fwn\") pod \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\" (UID: \"273ff4c8-4957-434f-a9f6-26ef0ed0f478\") " Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.232957 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/273ff4c8-4957-434f-a9f6-26ef0ed0f478-config-volume" (OuterVolumeSpecName: "config-volume") pod "273ff4c8-4957-434f-a9f6-26ef0ed0f478" (UID: "273ff4c8-4957-434f-a9f6-26ef0ed0f478"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.233561 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/273ff4c8-4957-434f-a9f6-26ef0ed0f478-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.241367 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/273ff4c8-4957-434f-a9f6-26ef0ed0f478-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "273ff4c8-4957-434f-a9f6-26ef0ed0f478" (UID: "273ff4c8-4957-434f-a9f6-26ef0ed0f478"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.241426 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/273ff4c8-4957-434f-a9f6-26ef0ed0f478-kube-api-access-q9fwn" (OuterVolumeSpecName: "kube-api-access-q9fwn") pod "273ff4c8-4957-434f-a9f6-26ef0ed0f478" (UID: "273ff4c8-4957-434f-a9f6-26ef0ed0f478"). InnerVolumeSpecName "kube-api-access-q9fwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.336020 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9fwn\" (UniqueName: \"kubernetes.io/projected/273ff4c8-4957-434f-a9f6-26ef0ed0f478-kube-api-access-q9fwn\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.336053 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/273ff4c8-4957-434f-a9f6-26ef0ed0f478-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.680834 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" event={"ID":"273ff4c8-4957-434f-a9f6-26ef0ed0f478","Type":"ContainerDied","Data":"95015be276e81c1a382def47c5869241cf8dee38c03886e51e2981791c45edf1"} Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.680876 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95015be276e81c1a382def47c5869241cf8dee38c03886e51e2981791c45edf1" Dec 05 13:30:03 crc kubenswrapper[4809]: I1205 13:30:03.680915 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415690-wx4h7" Dec 05 13:30:04 crc kubenswrapper[4809]: I1205 13:30:04.145292 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss"] Dec 05 13:30:04 crc kubenswrapper[4809]: I1205 13:30:04.153862 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415645-d42ss"] Dec 05 13:30:04 crc kubenswrapper[4809]: I1205 13:30:04.886317 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8912b56b-2f35-41b9-aa28-dba7a16846af" path="/var/lib/kubelet/pods/8912b56b-2f35-41b9-aa28-dba7a16846af/volumes" Dec 05 13:30:14 crc kubenswrapper[4809]: I1205 13:30:14.806818 4809 generic.go:334] "Generic (PLEG): container finished" podID="0fb44e23-96d1-4a58-b7df-0522db74bb22" containerID="5a8f598ffdb56508ad8725a6cf2bede703a2d436f46b7464fb64b391b4238a86" exitCode=0 Dec 05 13:30:14 crc kubenswrapper[4809]: I1205 13:30:14.806914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" event={"ID":"0fb44e23-96d1-4a58-b7df-0522db74bb22","Type":"ContainerDied","Data":"5a8f598ffdb56508ad8725a6cf2bede703a2d436f46b7464fb64b391b4238a86"} Dec 05 13:30:14 crc kubenswrapper[4809]: I1205 13:30:14.882980 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:30:14 crc kubenswrapper[4809]: E1205 13:30:14.883406 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.353343 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.445124 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgl7g\" (UniqueName: \"kubernetes.io/projected/0fb44e23-96d1-4a58-b7df-0522db74bb22-kube-api-access-tgl7g\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.445488 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-2\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.445613 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceph\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.445772 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-telemetry-combined-ca-bundle\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.445830 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-inventory\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.445889 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-1\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.446593 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ssh-key\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.446714 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-0\") pod \"0fb44e23-96d1-4a58-b7df-0522db74bb22\" (UID: \"0fb44e23-96d1-4a58-b7df-0522db74bb22\") " Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.451624 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceph" (OuterVolumeSpecName: "ceph") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.452357 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.452488 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fb44e23-96d1-4a58-b7df-0522db74bb22-kube-api-access-tgl7g" (OuterVolumeSpecName: "kube-api-access-tgl7g") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "kube-api-access-tgl7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.480041 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.483114 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.483843 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.488063 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-inventory" (OuterVolumeSpecName: "inventory") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.495837 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "0fb44e23-96d1-4a58-b7df-0522db74bb22" (UID: "0fb44e23-96d1-4a58-b7df-0522db74bb22"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550264 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550307 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550325 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550336 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550349 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgl7g\" (UniqueName: \"kubernetes.io/projected/0fb44e23-96d1-4a58-b7df-0522db74bb22-kube-api-access-tgl7g\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550364 4809 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550376 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.550387 4809 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb44e23-96d1-4a58-b7df-0522db74bb22-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.830883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" event={"ID":"0fb44e23-96d1-4a58-b7df-0522db74bb22","Type":"ContainerDied","Data":"341e14689a335ff8527d0aa878a2b640ef9aa9bcb0da603e5bb099ed53c5e656"} Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.830931 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="341e14689a335ff8527d0aa878a2b640ef9aa9bcb0da603e5bb099ed53c5e656" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.830969 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-7mb5p" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.955986 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x8pdw"] Dec 05 13:30:16 crc kubenswrapper[4809]: E1205 13:30:16.956766 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="273ff4c8-4957-434f-a9f6-26ef0ed0f478" containerName="collect-profiles" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.956809 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="273ff4c8-4957-434f-a9f6-26ef0ed0f478" containerName="collect-profiles" Dec 05 13:30:16 crc kubenswrapper[4809]: E1205 13:30:16.956852 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb44e23-96d1-4a58-b7df-0522db74bb22" containerName="telemetry-openstack-openstack-cell1" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.956870 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb44e23-96d1-4a58-b7df-0522db74bb22" containerName="telemetry-openstack-openstack-cell1" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.957430 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="273ff4c8-4957-434f-a9f6-26ef0ed0f478" containerName="collect-profiles" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.957517 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb44e23-96d1-4a58-b7df-0522db74bb22" containerName="telemetry-openstack-openstack-cell1" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.959060 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.963480 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.964611 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.964832 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.964903 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.971215 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:30:16 crc kubenswrapper[4809]: I1205 13:30:16.971518 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x8pdw"] Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.069591 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.070101 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.070276 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6bhc\" (UniqueName: \"kubernetes.io/projected/19eab448-bf57-4575-9a45-c50bf7b3fb77-kube-api-access-z6bhc\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.070384 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.070499 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.070561 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.172851 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.173224 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.173365 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6bhc\" (UniqueName: \"kubernetes.io/projected/19eab448-bf57-4575-9a45-c50bf7b3fb77-kube-api-access-z6bhc\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.173469 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.173577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.173690 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.177551 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.177947 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.179508 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.180079 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.185318 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.191723 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6bhc\" (UniqueName: \"kubernetes.io/projected/19eab448-bf57-4575-9a45-c50bf7b3fb77-kube-api-access-z6bhc\") pod \"neutron-sriov-openstack-openstack-cell1-x8pdw\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.298427 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:30:17 crc kubenswrapper[4809]: I1205 13:30:17.905138 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x8pdw"] Dec 05 13:30:18 crc kubenswrapper[4809]: I1205 13:30:18.869897 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" event={"ID":"19eab448-bf57-4575-9a45-c50bf7b3fb77","Type":"ContainerStarted","Data":"a6b54149e4519a8e44951cfe65381742b7afdfe790338ba67d8134258adc1822"} Dec 05 13:30:19 crc kubenswrapper[4809]: I1205 13:30:19.881704 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" event={"ID":"19eab448-bf57-4575-9a45-c50bf7b3fb77","Type":"ContainerStarted","Data":"5021cecc61b644a7c924bb3a466601a000639f98bf569c2590739321a7fc88fa"} Dec 05 13:30:19 crc kubenswrapper[4809]: I1205 13:30:19.904108 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" podStartSLOduration=2.615665065 podStartE2EDuration="3.904088156s" podCreationTimestamp="2025-12-05 13:30:16 +0000 UTC" firstStartedPulling="2025-12-05 13:30:17.912708342 +0000 UTC m=+8513.303684900" lastFinishedPulling="2025-12-05 13:30:19.201131413 +0000 UTC m=+8514.592107991" observedRunningTime="2025-12-05 13:30:19.897951183 +0000 UTC m=+8515.288927761" watchObservedRunningTime="2025-12-05 13:30:19.904088156 +0000 UTC m=+8515.295064714" Dec 05 13:30:25 crc kubenswrapper[4809]: I1205 13:30:25.872392 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:30:25 crc kubenswrapper[4809]: E1205 13:30:25.873313 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:30:38 crc kubenswrapper[4809]: I1205 13:30:38.873434 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:30:38 crc kubenswrapper[4809]: E1205 13:30:38.874229 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:30:47 crc kubenswrapper[4809]: I1205 13:30:47.054443 4809 scope.go:117] "RemoveContainer" containerID="ed3732aaa161829f7f1f0932a5fc022983f990af667f0a0a2d6883e075b39bb0" Dec 05 13:30:53 crc kubenswrapper[4809]: I1205 13:30:53.872144 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:30:53 crc kubenswrapper[4809]: E1205 13:30:53.872837 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:31:04 crc kubenswrapper[4809]: I1205 13:31:04.888580 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:31:04 crc kubenswrapper[4809]: E1205 13:31:04.889653 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:31:15 crc kubenswrapper[4809]: I1205 13:31:15.872869 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:31:15 crc kubenswrapper[4809]: E1205 13:31:15.874077 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:31:30 crc kubenswrapper[4809]: I1205 13:31:30.874327 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:31:30 crc kubenswrapper[4809]: E1205 13:31:30.875225 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:31:42 crc kubenswrapper[4809]: I1205 13:31:42.872399 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:31:42 crc kubenswrapper[4809]: E1205 13:31:42.873353 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:31:53 crc kubenswrapper[4809]: I1205 13:31:53.872525 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:31:53 crc kubenswrapper[4809]: E1205 13:31:53.873324 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:32:07 crc kubenswrapper[4809]: I1205 13:32:07.872438 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:32:07 crc kubenswrapper[4809]: E1205 13:32:07.873305 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:32:10 crc kubenswrapper[4809]: I1205 13:32:10.840420 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8bxsn"] Dec 05 13:32:10 crc kubenswrapper[4809]: I1205 13:32:10.844020 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:10 crc kubenswrapper[4809]: I1205 13:32:10.855604 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8bxsn"] Dec 05 13:32:10 crc kubenswrapper[4809]: I1205 13:32:10.920159 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-catalog-content\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:10 crc kubenswrapper[4809]: I1205 13:32:10.921179 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsxjd\" (UniqueName: \"kubernetes.io/projected/30256971-555a-4dc1-b8c5-c40bae999255-kube-api-access-lsxjd\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:10 crc kubenswrapper[4809]: I1205 13:32:10.921379 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-utilities\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.024459 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsxjd\" (UniqueName: \"kubernetes.io/projected/30256971-555a-4dc1-b8c5-c40bae999255-kube-api-access-lsxjd\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.024543 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-utilities\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.024766 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-catalog-content\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.025212 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-utilities\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.025303 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-catalog-content\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.045618 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsxjd\" (UniqueName: \"kubernetes.io/projected/30256971-555a-4dc1-b8c5-c40bae999255-kube-api-access-lsxjd\") pod \"certified-operators-8bxsn\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.183909 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:11 crc kubenswrapper[4809]: I1205 13:32:11.755085 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8bxsn"] Dec 05 13:32:12 crc kubenswrapper[4809]: I1205 13:32:12.106850 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bxsn" event={"ID":"30256971-555a-4dc1-b8c5-c40bae999255","Type":"ContainerStarted","Data":"344e29d2091dff9af3a95f4af0a4a3ac31d292783ad5504501139a2e29c02115"} Dec 05 13:32:13 crc kubenswrapper[4809]: I1205 13:32:13.134374 4809 generic.go:334] "Generic (PLEG): container finished" podID="30256971-555a-4dc1-b8c5-c40bae999255" containerID="3535280c6134b97e5ad058ca5692660bbb8a5c664ad31053db560b0c7cc3d003" exitCode=0 Dec 05 13:32:13 crc kubenswrapper[4809]: I1205 13:32:13.134443 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bxsn" event={"ID":"30256971-555a-4dc1-b8c5-c40bae999255","Type":"ContainerDied","Data":"3535280c6134b97e5ad058ca5692660bbb8a5c664ad31053db560b0c7cc3d003"} Dec 05 13:32:13 crc kubenswrapper[4809]: I1205 13:32:13.136693 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:32:14 crc kubenswrapper[4809]: I1205 13:32:14.145290 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bxsn" event={"ID":"30256971-555a-4dc1-b8c5-c40bae999255","Type":"ContainerStarted","Data":"da2acf373344d1d2239f27f2832c3df2c6bfbf7bcaff27f52e00546dc8cf917a"} Dec 05 13:32:15 crc kubenswrapper[4809]: I1205 13:32:15.179050 4809 generic.go:334] "Generic (PLEG): container finished" podID="30256971-555a-4dc1-b8c5-c40bae999255" containerID="da2acf373344d1d2239f27f2832c3df2c6bfbf7bcaff27f52e00546dc8cf917a" exitCode=0 Dec 05 13:32:15 crc kubenswrapper[4809]: I1205 13:32:15.179345 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bxsn" event={"ID":"30256971-555a-4dc1-b8c5-c40bae999255","Type":"ContainerDied","Data":"da2acf373344d1d2239f27f2832c3df2c6bfbf7bcaff27f52e00546dc8cf917a"} Dec 05 13:32:16 crc kubenswrapper[4809]: I1205 13:32:16.191774 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bxsn" event={"ID":"30256971-555a-4dc1-b8c5-c40bae999255","Type":"ContainerStarted","Data":"437283188ca0b77a8aec547279a75ac18f74539aef2dd2ce8d467c8911732eb9"} Dec 05 13:32:16 crc kubenswrapper[4809]: I1205 13:32:16.212272 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8bxsn" podStartSLOduration=3.7529137329999998 podStartE2EDuration="6.212236083s" podCreationTimestamp="2025-12-05 13:32:10 +0000 UTC" firstStartedPulling="2025-12-05 13:32:13.136394231 +0000 UTC m=+8628.527370789" lastFinishedPulling="2025-12-05 13:32:15.595716591 +0000 UTC m=+8630.986693139" observedRunningTime="2025-12-05 13:32:16.208445763 +0000 UTC m=+8631.599422341" watchObservedRunningTime="2025-12-05 13:32:16.212236083 +0000 UTC m=+8631.603212641" Dec 05 13:32:21 crc kubenswrapper[4809]: I1205 13:32:21.185422 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:21 crc kubenswrapper[4809]: I1205 13:32:21.185947 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:21 crc kubenswrapper[4809]: I1205 13:32:21.242225 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:21 crc kubenswrapper[4809]: I1205 13:32:21.315238 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:21 crc kubenswrapper[4809]: I1205 13:32:21.484241 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8bxsn"] Dec 05 13:32:21 crc kubenswrapper[4809]: I1205 13:32:21.872860 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:32:21 crc kubenswrapper[4809]: E1205 13:32:21.873168 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:32:23 crc kubenswrapper[4809]: I1205 13:32:23.283085 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8bxsn" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="registry-server" containerID="cri-o://437283188ca0b77a8aec547279a75ac18f74539aef2dd2ce8d467c8911732eb9" gracePeriod=2 Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.298351 4809 generic.go:334] "Generic (PLEG): container finished" podID="30256971-555a-4dc1-b8c5-c40bae999255" containerID="437283188ca0b77a8aec547279a75ac18f74539aef2dd2ce8d467c8911732eb9" exitCode=0 Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.298524 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bxsn" event={"ID":"30256971-555a-4dc1-b8c5-c40bae999255","Type":"ContainerDied","Data":"437283188ca0b77a8aec547279a75ac18f74539aef2dd2ce8d467c8911732eb9"} Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.652825 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.749300 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsxjd\" (UniqueName: \"kubernetes.io/projected/30256971-555a-4dc1-b8c5-c40bae999255-kube-api-access-lsxjd\") pod \"30256971-555a-4dc1-b8c5-c40bae999255\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.749381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-utilities\") pod \"30256971-555a-4dc1-b8c5-c40bae999255\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.749576 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-catalog-content\") pod \"30256971-555a-4dc1-b8c5-c40bae999255\" (UID: \"30256971-555a-4dc1-b8c5-c40bae999255\") " Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.750408 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-utilities" (OuterVolumeSpecName: "utilities") pod "30256971-555a-4dc1-b8c5-c40bae999255" (UID: "30256971-555a-4dc1-b8c5-c40bae999255"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.754593 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30256971-555a-4dc1-b8c5-c40bae999255-kube-api-access-lsxjd" (OuterVolumeSpecName: "kube-api-access-lsxjd") pod "30256971-555a-4dc1-b8c5-c40bae999255" (UID: "30256971-555a-4dc1-b8c5-c40bae999255"). InnerVolumeSpecName "kube-api-access-lsxjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.852024 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsxjd\" (UniqueName: \"kubernetes.io/projected/30256971-555a-4dc1-b8c5-c40bae999255-kube-api-access-lsxjd\") on node \"crc\" DevicePath \"\"" Dec 05 13:32:24 crc kubenswrapper[4809]: I1205 13:32:24.852059 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.316882 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8bxsn" event={"ID":"30256971-555a-4dc1-b8c5-c40bae999255","Type":"ContainerDied","Data":"344e29d2091dff9af3a95f4af0a4a3ac31d292783ad5504501139a2e29c02115"} Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.317033 4809 scope.go:117] "RemoveContainer" containerID="437283188ca0b77a8aec547279a75ac18f74539aef2dd2ce8d467c8911732eb9" Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.318913 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8bxsn" Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.354096 4809 scope.go:117] "RemoveContainer" containerID="da2acf373344d1d2239f27f2832c3df2c6bfbf7bcaff27f52e00546dc8cf917a" Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.396326 4809 scope.go:117] "RemoveContainer" containerID="3535280c6134b97e5ad058ca5692660bbb8a5c664ad31053db560b0c7cc3d003" Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.677161 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30256971-555a-4dc1-b8c5-c40bae999255" (UID: "30256971-555a-4dc1-b8c5-c40bae999255"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.773943 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30256971-555a-4dc1-b8c5-c40bae999255-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.954814 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8bxsn"] Dec 05 13:32:25 crc kubenswrapper[4809]: I1205 13:32:25.964070 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8bxsn"] Dec 05 13:32:26 crc kubenswrapper[4809]: I1205 13:32:26.890359 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30256971-555a-4dc1-b8c5-c40bae999255" path="/var/lib/kubelet/pods/30256971-555a-4dc1-b8c5-c40bae999255/volumes" Dec 05 13:32:34 crc kubenswrapper[4809]: I1205 13:32:34.880175 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:32:34 crc kubenswrapper[4809]: E1205 13:32:34.881520 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:32:47 crc kubenswrapper[4809]: I1205 13:32:47.872572 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:32:47 crc kubenswrapper[4809]: E1205 13:32:47.873459 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:32:58 crc kubenswrapper[4809]: I1205 13:32:58.872205 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:32:58 crc kubenswrapper[4809]: E1205 13:32:58.873119 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:33:09 crc kubenswrapper[4809]: I1205 13:33:09.872066 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:33:09 crc kubenswrapper[4809]: E1205 13:33:09.872778 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:33:21 crc kubenswrapper[4809]: I1205 13:33:21.873113 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:33:21 crc kubenswrapper[4809]: E1205 13:33:21.874534 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:33:35 crc kubenswrapper[4809]: I1205 13:33:35.872733 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:33:35 crc kubenswrapper[4809]: E1205 13:33:35.873765 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:33:47 crc kubenswrapper[4809]: I1205 13:33:47.871729 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:33:47 crc kubenswrapper[4809]: E1205 13:33:47.872548 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:33:52 crc kubenswrapper[4809]: I1205 13:33:52.378158 4809 generic.go:334] "Generic (PLEG): container finished" podID="19eab448-bf57-4575-9a45-c50bf7b3fb77" containerID="5021cecc61b644a7c924bb3a466601a000639f98bf569c2590739321a7fc88fa" exitCode=0 Dec 05 13:33:52 crc kubenswrapper[4809]: I1205 13:33:52.378256 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" event={"ID":"19eab448-bf57-4575-9a45-c50bf7b3fb77","Type":"ContainerDied","Data":"5021cecc61b644a7c924bb3a466601a000639f98bf569c2590739321a7fc88fa"} Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.005896 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.115070 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ssh-key\") pod \"19eab448-bf57-4575-9a45-c50bf7b3fb77\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.115122 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-agent-neutron-config-0\") pod \"19eab448-bf57-4575-9a45-c50bf7b3fb77\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.115293 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-combined-ca-bundle\") pod \"19eab448-bf57-4575-9a45-c50bf7b3fb77\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.115375 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6bhc\" (UniqueName: \"kubernetes.io/projected/19eab448-bf57-4575-9a45-c50bf7b3fb77-kube-api-access-z6bhc\") pod \"19eab448-bf57-4575-9a45-c50bf7b3fb77\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.115459 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ceph\") pod \"19eab448-bf57-4575-9a45-c50bf7b3fb77\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.115485 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-inventory\") pod \"19eab448-bf57-4575-9a45-c50bf7b3fb77\" (UID: \"19eab448-bf57-4575-9a45-c50bf7b3fb77\") " Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.120514 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ceph" (OuterVolumeSpecName: "ceph") pod "19eab448-bf57-4575-9a45-c50bf7b3fb77" (UID: "19eab448-bf57-4575-9a45-c50bf7b3fb77"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.124815 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19eab448-bf57-4575-9a45-c50bf7b3fb77-kube-api-access-z6bhc" (OuterVolumeSpecName: "kube-api-access-z6bhc") pod "19eab448-bf57-4575-9a45-c50bf7b3fb77" (UID: "19eab448-bf57-4575-9a45-c50bf7b3fb77"). InnerVolumeSpecName "kube-api-access-z6bhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.142105 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "19eab448-bf57-4575-9a45-c50bf7b3fb77" (UID: "19eab448-bf57-4575-9a45-c50bf7b3fb77"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.154977 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "19eab448-bf57-4575-9a45-c50bf7b3fb77" (UID: "19eab448-bf57-4575-9a45-c50bf7b3fb77"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.161186 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19eab448-bf57-4575-9a45-c50bf7b3fb77" (UID: "19eab448-bf57-4575-9a45-c50bf7b3fb77"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.163846 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-inventory" (OuterVolumeSpecName: "inventory") pod "19eab448-bf57-4575-9a45-c50bf7b3fb77" (UID: "19eab448-bf57-4575-9a45-c50bf7b3fb77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.218228 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.218270 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6bhc\" (UniqueName: \"kubernetes.io/projected/19eab448-bf57-4575-9a45-c50bf7b3fb77-kube-api-access-z6bhc\") on node \"crc\" DevicePath \"\"" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.218281 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.218290 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.218301 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.218310 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/19eab448-bf57-4575-9a45-c50bf7b3fb77-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.402804 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" event={"ID":"19eab448-bf57-4575-9a45-c50bf7b3fb77","Type":"ContainerDied","Data":"a6b54149e4519a8e44951cfe65381742b7afdfe790338ba67d8134258adc1822"} Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.402906 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6b54149e4519a8e44951cfe65381742b7afdfe790338ba67d8134258adc1822" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.402828 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x8pdw" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.511263 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq"] Dec 05 13:33:54 crc kubenswrapper[4809]: E1205 13:33:54.511703 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="extract-utilities" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.511715 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="extract-utilities" Dec 05 13:33:54 crc kubenswrapper[4809]: E1205 13:33:54.511730 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19eab448-bf57-4575-9a45-c50bf7b3fb77" containerName="neutron-sriov-openstack-openstack-cell1" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.511736 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="19eab448-bf57-4575-9a45-c50bf7b3fb77" containerName="neutron-sriov-openstack-openstack-cell1" Dec 05 13:33:54 crc kubenswrapper[4809]: E1205 13:33:54.511745 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="registry-server" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.511751 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="registry-server" Dec 05 13:33:54 crc kubenswrapper[4809]: E1205 13:33:54.511767 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="extract-content" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.511773 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="extract-content" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.512024 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="19eab448-bf57-4575-9a45-c50bf7b3fb77" containerName="neutron-sriov-openstack-openstack-cell1" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.512041 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="30256971-555a-4dc1-b8c5-c40bae999255" containerName="registry-server" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.512801 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.515467 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.515538 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.515827 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.516038 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.516185 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.550910 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq"] Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.615404 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-52ggh"] Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.618333 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.627967 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.628042 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.628070 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj4cn\" (UniqueName: \"kubernetes.io/projected/3b3d10d8-24e2-4f12-9007-3fd1368261c0-kube-api-access-sj4cn\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.628104 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.628130 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.628401 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.633877 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52ggh"] Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.730977 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731053 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj4cn\" (UniqueName: \"kubernetes.io/projected/3b3d10d8-24e2-4f12-9007-3fd1368261c0-kube-api-access-sj4cn\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731102 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-utilities\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731144 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731185 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731220 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-catalog-content\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731245 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27dtd\" (UniqueName: \"kubernetes.io/projected/b32d60a7-b7b4-4a6d-829d-58af734b0506-kube-api-access-27dtd\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731292 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.731413 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.736866 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.738118 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.739094 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.739506 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.741183 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.756728 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj4cn\" (UniqueName: \"kubernetes.io/projected/3b3d10d8-24e2-4f12-9007-3fd1368261c0-kube-api-access-sj4cn\") pod \"neutron-dhcp-openstack-openstack-cell1-qd9cq\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.832760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-catalog-content\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.832825 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27dtd\" (UniqueName: \"kubernetes.io/projected/b32d60a7-b7b4-4a6d-829d-58af734b0506-kube-api-access-27dtd\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.833094 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-utilities\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.833724 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-utilities\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.833981 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-catalog-content\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.834362 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.854344 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27dtd\" (UniqueName: \"kubernetes.io/projected/b32d60a7-b7b4-4a6d-829d-58af734b0506-kube-api-access-27dtd\") pod \"redhat-marketplace-52ggh\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:54 crc kubenswrapper[4809]: I1205 13:33:54.962134 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:33:55 crc kubenswrapper[4809]: I1205 13:33:55.506148 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq"] Dec 05 13:33:55 crc kubenswrapper[4809]: W1205 13:33:55.550252 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb32d60a7_b7b4_4a6d_829d_58af734b0506.slice/crio-116da8e439e02b65cdd349faaddda4ea838f311dde98289e657e58a7075cdfd9 WatchSource:0}: Error finding container 116da8e439e02b65cdd349faaddda4ea838f311dde98289e657e58a7075cdfd9: Status 404 returned error can't find the container with id 116da8e439e02b65cdd349faaddda4ea838f311dde98289e657e58a7075cdfd9 Dec 05 13:33:55 crc kubenswrapper[4809]: I1205 13:33:55.553927 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52ggh"] Dec 05 13:33:56 crc kubenswrapper[4809]: I1205 13:33:56.421517 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" event={"ID":"3b3d10d8-24e2-4f12-9007-3fd1368261c0","Type":"ContainerStarted","Data":"40d9dd8eb1eb0a01ea17a471771a226bf2445e2708def9ac0915bd9bb0a01bdd"} Dec 05 13:33:56 crc kubenswrapper[4809]: I1205 13:33:56.423500 4809 generic.go:334] "Generic (PLEG): container finished" podID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerID="199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244" exitCode=0 Dec 05 13:33:56 crc kubenswrapper[4809]: I1205 13:33:56.423534 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52ggh" event={"ID":"b32d60a7-b7b4-4a6d-829d-58af734b0506","Type":"ContainerDied","Data":"199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244"} Dec 05 13:33:56 crc kubenswrapper[4809]: I1205 13:33:56.423553 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52ggh" event={"ID":"b32d60a7-b7b4-4a6d-829d-58af734b0506","Type":"ContainerStarted","Data":"116da8e439e02b65cdd349faaddda4ea838f311dde98289e657e58a7075cdfd9"} Dec 05 13:33:57 crc kubenswrapper[4809]: I1205 13:33:57.434165 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" event={"ID":"3b3d10d8-24e2-4f12-9007-3fd1368261c0","Type":"ContainerStarted","Data":"5c9502862f8061cb3d8144b5b6b1f1b7f2f30af93b276d31847c659961cc4c92"} Dec 05 13:33:57 crc kubenswrapper[4809]: I1205 13:33:57.462064 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" podStartSLOduration=2.624902483 podStartE2EDuration="3.4620413s" podCreationTimestamp="2025-12-05 13:33:54 +0000 UTC" firstStartedPulling="2025-12-05 13:33:55.516750287 +0000 UTC m=+8730.907726845" lastFinishedPulling="2025-12-05 13:33:56.353889104 +0000 UTC m=+8731.744865662" observedRunningTime="2025-12-05 13:33:57.449827596 +0000 UTC m=+8732.840804154" watchObservedRunningTime="2025-12-05 13:33:57.4620413 +0000 UTC m=+8732.853017868" Dec 05 13:34:00 crc kubenswrapper[4809]: I1205 13:34:00.466268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52ggh" event={"ID":"b32d60a7-b7b4-4a6d-829d-58af734b0506","Type":"ContainerStarted","Data":"7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0"} Dec 05 13:34:01 crc kubenswrapper[4809]: I1205 13:34:01.482695 4809 generic.go:334] "Generic (PLEG): container finished" podID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerID="7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0" exitCode=0 Dec 05 13:34:01 crc kubenswrapper[4809]: I1205 13:34:01.482786 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52ggh" event={"ID":"b32d60a7-b7b4-4a6d-829d-58af734b0506","Type":"ContainerDied","Data":"7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0"} Dec 05 13:34:01 crc kubenswrapper[4809]: I1205 13:34:01.873298 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:34:01 crc kubenswrapper[4809]: E1205 13:34:01.873608 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:34:03 crc kubenswrapper[4809]: I1205 13:34:03.503344 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52ggh" event={"ID":"b32d60a7-b7b4-4a6d-829d-58af734b0506","Type":"ContainerStarted","Data":"c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072"} Dec 05 13:34:03 crc kubenswrapper[4809]: I1205 13:34:03.540594 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-52ggh" podStartSLOduration=3.227095604 podStartE2EDuration="9.540567882s" podCreationTimestamp="2025-12-05 13:33:54 +0000 UTC" firstStartedPulling="2025-12-05 13:33:56.425831199 +0000 UTC m=+8731.816807757" lastFinishedPulling="2025-12-05 13:34:02.739303477 +0000 UTC m=+8738.130280035" observedRunningTime="2025-12-05 13:34:03.530560475 +0000 UTC m=+8738.921537053" watchObservedRunningTime="2025-12-05 13:34:03.540567882 +0000 UTC m=+8738.931544440" Dec 05 13:34:04 crc kubenswrapper[4809]: I1205 13:34:04.962578 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:34:04 crc kubenswrapper[4809]: I1205 13:34:04.963017 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:34:05 crc kubenswrapper[4809]: I1205 13:34:05.039516 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:34:14 crc kubenswrapper[4809]: I1205 13:34:14.881381 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:34:14 crc kubenswrapper[4809]: E1205 13:34:14.882318 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:34:15 crc kubenswrapper[4809]: I1205 13:34:15.017429 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:34:15 crc kubenswrapper[4809]: I1205 13:34:15.073587 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52ggh"] Dec 05 13:34:15 crc kubenswrapper[4809]: I1205 13:34:15.633189 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-52ggh" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="registry-server" containerID="cri-o://c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072" gracePeriod=2 Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.349242 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.443842 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-utilities\") pod \"b32d60a7-b7b4-4a6d-829d-58af734b0506\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.443939 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27dtd\" (UniqueName: \"kubernetes.io/projected/b32d60a7-b7b4-4a6d-829d-58af734b0506-kube-api-access-27dtd\") pod \"b32d60a7-b7b4-4a6d-829d-58af734b0506\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.444178 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-catalog-content\") pod \"b32d60a7-b7b4-4a6d-829d-58af734b0506\" (UID: \"b32d60a7-b7b4-4a6d-829d-58af734b0506\") " Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.444698 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-utilities" (OuterVolumeSpecName: "utilities") pod "b32d60a7-b7b4-4a6d-829d-58af734b0506" (UID: "b32d60a7-b7b4-4a6d-829d-58af734b0506"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.445326 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.466524 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b32d60a7-b7b4-4a6d-829d-58af734b0506" (UID: "b32d60a7-b7b4-4a6d-829d-58af734b0506"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.547089 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b32d60a7-b7b4-4a6d-829d-58af734b0506-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.645463 4809 generic.go:334] "Generic (PLEG): container finished" podID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerID="c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072" exitCode=0 Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.645503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52ggh" event={"ID":"b32d60a7-b7b4-4a6d-829d-58af734b0506","Type":"ContainerDied","Data":"c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072"} Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.645616 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52ggh" event={"ID":"b32d60a7-b7b4-4a6d-829d-58af734b0506","Type":"ContainerDied","Data":"116da8e439e02b65cdd349faaddda4ea838f311dde98289e657e58a7075cdfd9"} Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.645676 4809 scope.go:117] "RemoveContainer" containerID="c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072" Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.646287 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52ggh" Dec 05 13:34:16 crc kubenswrapper[4809]: I1205 13:34:16.671611 4809 scope.go:117] "RemoveContainer" containerID="7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.086070 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b32d60a7-b7b4-4a6d-829d-58af734b0506-kube-api-access-27dtd" (OuterVolumeSpecName: "kube-api-access-27dtd") pod "b32d60a7-b7b4-4a6d-829d-58af734b0506" (UID: "b32d60a7-b7b4-4a6d-829d-58af734b0506"). InnerVolumeSpecName "kube-api-access-27dtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.107830 4809 scope.go:117] "RemoveContainer" containerID="199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.162955 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27dtd\" (UniqueName: \"kubernetes.io/projected/b32d60a7-b7b4-4a6d-829d-58af734b0506-kube-api-access-27dtd\") on node \"crc\" DevicePath \"\"" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.218656 4809 scope.go:117] "RemoveContainer" containerID="c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072" Dec 05 13:34:17 crc kubenswrapper[4809]: E1205 13:34:17.219117 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072\": container with ID starting with c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072 not found: ID does not exist" containerID="c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.219155 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072"} err="failed to get container status \"c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072\": rpc error: code = NotFound desc = could not find container \"c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072\": container with ID starting with c4fa34cc15a1e8c2cbfabc95b0ed353a2f5f335634a64145735edf79dd405072 not found: ID does not exist" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.219184 4809 scope.go:117] "RemoveContainer" containerID="7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0" Dec 05 13:34:17 crc kubenswrapper[4809]: E1205 13:34:17.219386 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0\": container with ID starting with 7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0 not found: ID does not exist" containerID="7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.219409 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0"} err="failed to get container status \"7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0\": rpc error: code = NotFound desc = could not find container \"7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0\": container with ID starting with 7da4eb31d15d348e705eb7590d455fa286f18c771b62d3351a9125d016d3b9b0 not found: ID does not exist" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.219422 4809 scope.go:117] "RemoveContainer" containerID="199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244" Dec 05 13:34:17 crc kubenswrapper[4809]: E1205 13:34:17.219700 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244\": container with ID starting with 199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244 not found: ID does not exist" containerID="199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.219725 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244"} err="failed to get container status \"199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244\": rpc error: code = NotFound desc = could not find container \"199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244\": container with ID starting with 199e7581894f0e3257b3bc8f79caffa8507f7a36e29900e9d60b1d1df5a7a244 not found: ID does not exist" Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.290036 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52ggh"] Dec 05 13:34:17 crc kubenswrapper[4809]: I1205 13:34:17.301953 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-52ggh"] Dec 05 13:34:18 crc kubenswrapper[4809]: I1205 13:34:18.884222 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" path="/var/lib/kubelet/pods/b32d60a7-b7b4-4a6d-829d-58af734b0506/volumes" Dec 05 13:34:28 crc kubenswrapper[4809]: I1205 13:34:28.872669 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:34:28 crc kubenswrapper[4809]: E1205 13:34:28.873531 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:34:42 crc kubenswrapper[4809]: I1205 13:34:42.874206 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:34:42 crc kubenswrapper[4809]: E1205 13:34:42.875459 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:34:55 crc kubenswrapper[4809]: I1205 13:34:55.873094 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:34:57 crc kubenswrapper[4809]: I1205 13:34:57.090200 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"ee2b1b16721c3b6a23331471a609811f8cfbcb2367cda45319070a7cccde75a2"} Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.344831 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8w7gm"] Dec 05 13:37:04 crc kubenswrapper[4809]: E1205 13:37:04.346234 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="extract-content" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.346252 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="extract-content" Dec 05 13:37:04 crc kubenswrapper[4809]: E1205 13:37:04.346285 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="extract-utilities" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.346293 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="extract-utilities" Dec 05 13:37:04 crc kubenswrapper[4809]: E1205 13:37:04.346335 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="registry-server" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.346345 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="registry-server" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.346657 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b32d60a7-b7b4-4a6d-829d-58af734b0506" containerName="registry-server" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.348560 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.363185 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8w7gm"] Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.448960 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-utilities\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.449055 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-catalog-content\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.449111 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db6kl\" (UniqueName: \"kubernetes.io/projected/456ff3a9-8ea2-42f1-addd-4c030b77fb20-kube-api-access-db6kl\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.551032 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-utilities\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.551236 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-catalog-content\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.551343 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db6kl\" (UniqueName: \"kubernetes.io/projected/456ff3a9-8ea2-42f1-addd-4c030b77fb20-kube-api-access-db6kl\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.551684 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-utilities\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.551731 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-catalog-content\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.572845 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db6kl\" (UniqueName: \"kubernetes.io/projected/456ff3a9-8ea2-42f1-addd-4c030b77fb20-kube-api-access-db6kl\") pod \"redhat-operators-8w7gm\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:04 crc kubenswrapper[4809]: I1205 13:37:04.676581 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:05 crc kubenswrapper[4809]: I1205 13:37:05.674244 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8w7gm"] Dec 05 13:37:06 crc kubenswrapper[4809]: I1205 13:37:06.477229 4809 generic.go:334] "Generic (PLEG): container finished" podID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerID="0be806feae39081051529d3d435353e086d03a85766f19b29677649e74a52e2e" exitCode=0 Dec 05 13:37:06 crc kubenswrapper[4809]: I1205 13:37:06.479753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8w7gm" event={"ID":"456ff3a9-8ea2-42f1-addd-4c030b77fb20","Type":"ContainerDied","Data":"0be806feae39081051529d3d435353e086d03a85766f19b29677649e74a52e2e"} Dec 05 13:37:06 crc kubenswrapper[4809]: I1205 13:37:06.479874 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8w7gm" event={"ID":"456ff3a9-8ea2-42f1-addd-4c030b77fb20","Type":"ContainerStarted","Data":"566ff2ff751ce15343ed667b27fae1050b4ab71d1a5b3592daab477b1061dcba"} Dec 05 13:37:07 crc kubenswrapper[4809]: I1205 13:37:07.490407 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8w7gm" event={"ID":"456ff3a9-8ea2-42f1-addd-4c030b77fb20","Type":"ContainerStarted","Data":"d53500f0846bc15721614b742796cfc8f5f3052be546dd1b3206484479a63386"} Dec 05 13:37:11 crc kubenswrapper[4809]: I1205 13:37:11.531756 4809 generic.go:334] "Generic (PLEG): container finished" podID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerID="d53500f0846bc15721614b742796cfc8f5f3052be546dd1b3206484479a63386" exitCode=0 Dec 05 13:37:11 crc kubenswrapper[4809]: I1205 13:37:11.532311 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8w7gm" event={"ID":"456ff3a9-8ea2-42f1-addd-4c030b77fb20","Type":"ContainerDied","Data":"d53500f0846bc15721614b742796cfc8f5f3052be546dd1b3206484479a63386"} Dec 05 13:37:13 crc kubenswrapper[4809]: I1205 13:37:13.560969 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8w7gm" event={"ID":"456ff3a9-8ea2-42f1-addd-4c030b77fb20","Type":"ContainerStarted","Data":"6c5f765f59f66dddb4c2eed67bce12a559b4d6afa312bb3effbc96c0111b992d"} Dec 05 13:37:14 crc kubenswrapper[4809]: I1205 13:37:14.046766 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:37:14 crc kubenswrapper[4809]: I1205 13:37:14.046875 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:37:14 crc kubenswrapper[4809]: I1205 13:37:14.676790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:14 crc kubenswrapper[4809]: I1205 13:37:14.676840 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:15 crc kubenswrapper[4809]: I1205 13:37:15.725338 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8w7gm" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="registry-server" probeResult="failure" output=< Dec 05 13:37:15 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 13:37:15 crc kubenswrapper[4809]: > Dec 05 13:37:24 crc kubenswrapper[4809]: I1205 13:37:24.745108 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:24 crc kubenswrapper[4809]: I1205 13:37:24.767570 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8w7gm" podStartSLOduration=14.265134071 podStartE2EDuration="20.76754698s" podCreationTimestamp="2025-12-05 13:37:04 +0000 UTC" firstStartedPulling="2025-12-05 13:37:06.495832146 +0000 UTC m=+8921.886808714" lastFinishedPulling="2025-12-05 13:37:12.998245055 +0000 UTC m=+8928.389221623" observedRunningTime="2025-12-05 13:37:13.583870155 +0000 UTC m=+8928.974846723" watchObservedRunningTime="2025-12-05 13:37:24.76754698 +0000 UTC m=+8940.158523538" Dec 05 13:37:24 crc kubenswrapper[4809]: I1205 13:37:24.797253 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:24 crc kubenswrapper[4809]: I1205 13:37:24.984355 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8w7gm"] Dec 05 13:37:26 crc kubenswrapper[4809]: I1205 13:37:26.701204 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8w7gm" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="registry-server" containerID="cri-o://6c5f765f59f66dddb4c2eed67bce12a559b4d6afa312bb3effbc96c0111b992d" gracePeriod=2 Dec 05 13:37:27 crc kubenswrapper[4809]: I1205 13:37:27.715907 4809 generic.go:334] "Generic (PLEG): container finished" podID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerID="6c5f765f59f66dddb4c2eed67bce12a559b4d6afa312bb3effbc96c0111b992d" exitCode=0 Dec 05 13:37:27 crc kubenswrapper[4809]: I1205 13:37:27.715966 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8w7gm" event={"ID":"456ff3a9-8ea2-42f1-addd-4c030b77fb20","Type":"ContainerDied","Data":"6c5f765f59f66dddb4c2eed67bce12a559b4d6afa312bb3effbc96c0111b992d"} Dec 05 13:37:27 crc kubenswrapper[4809]: I1205 13:37:27.716454 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8w7gm" event={"ID":"456ff3a9-8ea2-42f1-addd-4c030b77fb20","Type":"ContainerDied","Data":"566ff2ff751ce15343ed667b27fae1050b4ab71d1a5b3592daab477b1061dcba"} Dec 05 13:37:27 crc kubenswrapper[4809]: I1205 13:37:27.716486 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="566ff2ff751ce15343ed667b27fae1050b4ab71d1a5b3592daab477b1061dcba" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.535550 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.588674 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-utilities\") pod \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.588789 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-catalog-content\") pod \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.588825 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db6kl\" (UniqueName: \"kubernetes.io/projected/456ff3a9-8ea2-42f1-addd-4c030b77fb20-kube-api-access-db6kl\") pod \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\" (UID: \"456ff3a9-8ea2-42f1-addd-4c030b77fb20\") " Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.595804 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-utilities" (OuterVolumeSpecName: "utilities") pod "456ff3a9-8ea2-42f1-addd-4c030b77fb20" (UID: "456ff3a9-8ea2-42f1-addd-4c030b77fb20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.603562 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/456ff3a9-8ea2-42f1-addd-4c030b77fb20-kube-api-access-db6kl" (OuterVolumeSpecName: "kube-api-access-db6kl") pod "456ff3a9-8ea2-42f1-addd-4c030b77fb20" (UID: "456ff3a9-8ea2-42f1-addd-4c030b77fb20"). InnerVolumeSpecName "kube-api-access-db6kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.692182 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.692235 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db6kl\" (UniqueName: \"kubernetes.io/projected/456ff3a9-8ea2-42f1-addd-4c030b77fb20-kube-api-access-db6kl\") on node \"crc\" DevicePath \"\"" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.708796 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "456ff3a9-8ea2-42f1-addd-4c030b77fb20" (UID: "456ff3a9-8ea2-42f1-addd-4c030b77fb20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.726366 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8w7gm" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.778165 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8w7gm"] Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.797662 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456ff3a9-8ea2-42f1-addd-4c030b77fb20-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.809477 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8w7gm"] Dec 05 13:37:28 crc kubenswrapper[4809]: I1205 13:37:28.885112 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" path="/var/lib/kubelet/pods/456ff3a9-8ea2-42f1-addd-4c030b77fb20/volumes" Dec 05 13:37:44 crc kubenswrapper[4809]: I1205 13:37:44.046594 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:37:44 crc kubenswrapper[4809]: I1205 13:37:44.047538 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.046757 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.047563 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.047689 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.049005 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee2b1b16721c3b6a23331471a609811f8cfbcb2367cda45319070a7cccde75a2"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.049177 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://ee2b1b16721c3b6a23331471a609811f8cfbcb2367cda45319070a7cccde75a2" gracePeriod=600 Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.207609 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="ee2b1b16721c3b6a23331471a609811f8cfbcb2367cda45319070a7cccde75a2" exitCode=0 Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.207745 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"ee2b1b16721c3b6a23331471a609811f8cfbcb2367cda45319070a7cccde75a2"} Dec 05 13:38:14 crc kubenswrapper[4809]: I1205 13:38:14.207800 4809 scope.go:117] "RemoveContainer" containerID="f55c51490f384d1765bfba04954d7fd76f7e78538fccae56987658f2ef85e50c" Dec 05 13:38:15 crc kubenswrapper[4809]: I1205 13:38:15.221231 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103"} Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.115018 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2sjv7"] Dec 05 13:38:31 crc kubenswrapper[4809]: E1205 13:38:31.116069 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="extract-content" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.116085 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="extract-content" Dec 05 13:38:31 crc kubenswrapper[4809]: E1205 13:38:31.116103 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="extract-utilities" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.116110 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="extract-utilities" Dec 05 13:38:31 crc kubenswrapper[4809]: E1205 13:38:31.116123 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="registry-server" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.116131 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="registry-server" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.116364 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="456ff3a9-8ea2-42f1-addd-4c030b77fb20" containerName="registry-server" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.118281 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.138191 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2sjv7"] Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.206676 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqcwb\" (UniqueName: \"kubernetes.io/projected/d60bbc52-3283-4147-8ad8-28f9d0bb9281-kube-api-access-zqcwb\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.206757 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-utilities\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.206975 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-catalog-content\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.308587 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-utilities\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.308684 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-catalog-content\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.308827 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqcwb\" (UniqueName: \"kubernetes.io/projected/d60bbc52-3283-4147-8ad8-28f9d0bb9281-kube-api-access-zqcwb\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.309103 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-utilities\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.309410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-catalog-content\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.335479 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqcwb\" (UniqueName: \"kubernetes.io/projected/d60bbc52-3283-4147-8ad8-28f9d0bb9281-kube-api-access-zqcwb\") pod \"community-operators-2sjv7\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.441971 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:31 crc kubenswrapper[4809]: I1205 13:38:31.979300 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2sjv7"] Dec 05 13:38:33 crc kubenswrapper[4809]: I1205 13:38:33.421186 4809 generic.go:334] "Generic (PLEG): container finished" podID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerID="64c59db25bdb793aa93edb419a273c5d4474971c77106c436efe322f56a6c2e7" exitCode=0 Dec 05 13:38:33 crc kubenswrapper[4809]: I1205 13:38:33.421286 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sjv7" event={"ID":"d60bbc52-3283-4147-8ad8-28f9d0bb9281","Type":"ContainerDied","Data":"64c59db25bdb793aa93edb419a273c5d4474971c77106c436efe322f56a6c2e7"} Dec 05 13:38:33 crc kubenswrapper[4809]: I1205 13:38:33.422788 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sjv7" event={"ID":"d60bbc52-3283-4147-8ad8-28f9d0bb9281","Type":"ContainerStarted","Data":"ccce681633d01fcc2addab7cf516a772fdb54e464b8b86d7b31a187b9cf49c3f"} Dec 05 13:38:33 crc kubenswrapper[4809]: I1205 13:38:33.424454 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:38:35 crc kubenswrapper[4809]: I1205 13:38:35.447441 4809 generic.go:334] "Generic (PLEG): container finished" podID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerID="511a4ad5ed3a430abe256820a7d4cf0c322b91c870349ba77ac8121a461f03ff" exitCode=0 Dec 05 13:38:35 crc kubenswrapper[4809]: I1205 13:38:35.447503 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sjv7" event={"ID":"d60bbc52-3283-4147-8ad8-28f9d0bb9281","Type":"ContainerDied","Data":"511a4ad5ed3a430abe256820a7d4cf0c322b91c870349ba77ac8121a461f03ff"} Dec 05 13:38:43 crc kubenswrapper[4809]: I1205 13:38:43.537146 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sjv7" event={"ID":"d60bbc52-3283-4147-8ad8-28f9d0bb9281","Type":"ContainerStarted","Data":"9836e72459ede30b71f48a32168b7bac9a852ae40df62d056ea730659757e2ae"} Dec 05 13:38:43 crc kubenswrapper[4809]: I1205 13:38:43.569784 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2sjv7" podStartSLOduration=4.08366651 podStartE2EDuration="12.569755531s" podCreationTimestamp="2025-12-05 13:38:31 +0000 UTC" firstStartedPulling="2025-12-05 13:38:33.424200961 +0000 UTC m=+9008.815177519" lastFinishedPulling="2025-12-05 13:38:41.910289982 +0000 UTC m=+9017.301266540" observedRunningTime="2025-12-05 13:38:43.554439483 +0000 UTC m=+9018.945416081" watchObservedRunningTime="2025-12-05 13:38:43.569755531 +0000 UTC m=+9018.960732129" Dec 05 13:38:51 crc kubenswrapper[4809]: I1205 13:38:51.442607 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:51 crc kubenswrapper[4809]: I1205 13:38:51.443251 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:51 crc kubenswrapper[4809]: I1205 13:38:51.531191 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:51 crc kubenswrapper[4809]: I1205 13:38:51.668483 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:51 crc kubenswrapper[4809]: I1205 13:38:51.820806 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2sjv7"] Dec 05 13:38:53 crc kubenswrapper[4809]: I1205 13:38:53.652062 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2sjv7" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="registry-server" containerID="cri-o://9836e72459ede30b71f48a32168b7bac9a852ae40df62d056ea730659757e2ae" gracePeriod=2 Dec 05 13:38:54 crc kubenswrapper[4809]: I1205 13:38:54.663264 4809 generic.go:334] "Generic (PLEG): container finished" podID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerID="9836e72459ede30b71f48a32168b7bac9a852ae40df62d056ea730659757e2ae" exitCode=0 Dec 05 13:38:54 crc kubenswrapper[4809]: I1205 13:38:54.663835 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sjv7" event={"ID":"d60bbc52-3283-4147-8ad8-28f9d0bb9281","Type":"ContainerDied","Data":"9836e72459ede30b71f48a32168b7bac9a852ae40df62d056ea730659757e2ae"} Dec 05 13:38:54 crc kubenswrapper[4809]: I1205 13:38:54.883383 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:54 crc kubenswrapper[4809]: I1205 13:38:54.944550 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-catalog-content\") pod \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " Dec 05 13:38:54 crc kubenswrapper[4809]: I1205 13:38:54.944829 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-utilities\") pod \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " Dec 05 13:38:54 crc kubenswrapper[4809]: I1205 13:38:54.944945 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqcwb\" (UniqueName: \"kubernetes.io/projected/d60bbc52-3283-4147-8ad8-28f9d0bb9281-kube-api-access-zqcwb\") pod \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\" (UID: \"d60bbc52-3283-4147-8ad8-28f9d0bb9281\") " Dec 05 13:38:54 crc kubenswrapper[4809]: I1205 13:38:54.945784 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-utilities" (OuterVolumeSpecName: "utilities") pod "d60bbc52-3283-4147-8ad8-28f9d0bb9281" (UID: "d60bbc52-3283-4147-8ad8-28f9d0bb9281"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:54.996040 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d60bbc52-3283-4147-8ad8-28f9d0bb9281" (UID: "d60bbc52-3283-4147-8ad8-28f9d0bb9281"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.047019 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.047339 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d60bbc52-3283-4147-8ad8-28f9d0bb9281-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.105355 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d60bbc52-3283-4147-8ad8-28f9d0bb9281-kube-api-access-zqcwb" (OuterVolumeSpecName: "kube-api-access-zqcwb") pod "d60bbc52-3283-4147-8ad8-28f9d0bb9281" (UID: "d60bbc52-3283-4147-8ad8-28f9d0bb9281"). InnerVolumeSpecName "kube-api-access-zqcwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.149053 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqcwb\" (UniqueName: \"kubernetes.io/projected/d60bbc52-3283-4147-8ad8-28f9d0bb9281-kube-api-access-zqcwb\") on node \"crc\" DevicePath \"\"" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.679883 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2sjv7" event={"ID":"d60bbc52-3283-4147-8ad8-28f9d0bb9281","Type":"ContainerDied","Data":"ccce681633d01fcc2addab7cf516a772fdb54e464b8b86d7b31a187b9cf49c3f"} Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.679941 4809 scope.go:117] "RemoveContainer" containerID="9836e72459ede30b71f48a32168b7bac9a852ae40df62d056ea730659757e2ae" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.680021 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2sjv7" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.713591 4809 scope.go:117] "RemoveContainer" containerID="511a4ad5ed3a430abe256820a7d4cf0c322b91c870349ba77ac8121a461f03ff" Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.746203 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2sjv7"] Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.759964 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2sjv7"] Dec 05 13:38:55 crc kubenswrapper[4809]: I1205 13:38:55.761513 4809 scope.go:117] "RemoveContainer" containerID="64c59db25bdb793aa93edb419a273c5d4474971c77106c436efe322f56a6c2e7" Dec 05 13:38:56 crc kubenswrapper[4809]: I1205 13:38:56.885662 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" path="/var/lib/kubelet/pods/d60bbc52-3283-4147-8ad8-28f9d0bb9281/volumes" Dec 05 13:40:02 crc kubenswrapper[4809]: I1205 13:40:02.425593 4809 generic.go:334] "Generic (PLEG): container finished" podID="3b3d10d8-24e2-4f12-9007-3fd1368261c0" containerID="5c9502862f8061cb3d8144b5b6b1f1b7f2f30af93b276d31847c659961cc4c92" exitCode=0 Dec 05 13:40:02 crc kubenswrapper[4809]: I1205 13:40:02.425667 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" event={"ID":"3b3d10d8-24e2-4f12-9007-3fd1368261c0","Type":"ContainerDied","Data":"5c9502862f8061cb3d8144b5b6b1f1b7f2f30af93b276d31847c659961cc4c92"} Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.597646 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.723086 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-agent-neutron-config-0\") pod \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.723519 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-inventory\") pod \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.723915 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj4cn\" (UniqueName: \"kubernetes.io/projected/3b3d10d8-24e2-4f12-9007-3fd1368261c0-kube-api-access-sj4cn\") pod \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.724794 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ceph\") pod \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.725004 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-combined-ca-bundle\") pod \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.725172 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ssh-key\") pod \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\" (UID: \"3b3d10d8-24e2-4f12-9007-3fd1368261c0\") " Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.728744 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ceph" (OuterVolumeSpecName: "ceph") pod "3b3d10d8-24e2-4f12-9007-3fd1368261c0" (UID: "3b3d10d8-24e2-4f12-9007-3fd1368261c0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.729106 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "3b3d10d8-24e2-4f12-9007-3fd1368261c0" (UID: "3b3d10d8-24e2-4f12-9007-3fd1368261c0"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.737700 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b3d10d8-24e2-4f12-9007-3fd1368261c0-kube-api-access-sj4cn" (OuterVolumeSpecName: "kube-api-access-sj4cn") pod "3b3d10d8-24e2-4f12-9007-3fd1368261c0" (UID: "3b3d10d8-24e2-4f12-9007-3fd1368261c0"). InnerVolumeSpecName "kube-api-access-sj4cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.751704 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-inventory" (OuterVolumeSpecName: "inventory") pod "3b3d10d8-24e2-4f12-9007-3fd1368261c0" (UID: "3b3d10d8-24e2-4f12-9007-3fd1368261c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.755883 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b3d10d8-24e2-4f12-9007-3fd1368261c0" (UID: "3b3d10d8-24e2-4f12-9007-3fd1368261c0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.756400 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "3b3d10d8-24e2-4f12-9007-3fd1368261c0" (UID: "3b3d10d8-24e2-4f12-9007-3fd1368261c0"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.828766 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.828811 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.828825 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj4cn\" (UniqueName: \"kubernetes.io/projected/3b3d10d8-24e2-4f12-9007-3fd1368261c0-kube-api-access-sj4cn\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.828834 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.828843 4809 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:04 crc kubenswrapper[4809]: I1205 13:40:04.828855 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b3d10d8-24e2-4f12-9007-3fd1368261c0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:05 crc kubenswrapper[4809]: I1205 13:40:05.456357 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" event={"ID":"3b3d10d8-24e2-4f12-9007-3fd1368261c0","Type":"ContainerDied","Data":"40d9dd8eb1eb0a01ea17a471771a226bf2445e2708def9ac0915bd9bb0a01bdd"} Dec 05 13:40:05 crc kubenswrapper[4809]: I1205 13:40:05.456611 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40d9dd8eb1eb0a01ea17a471771a226bf2445e2708def9ac0915bd9bb0a01bdd" Dec 05 13:40:05 crc kubenswrapper[4809]: I1205 13:40:05.456447 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qd9cq" Dec 05 13:40:12 crc kubenswrapper[4809]: I1205 13:40:12.431588 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 13:40:12 crc kubenswrapper[4809]: I1205 13:40:12.432355 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="daf4bee5-c20a-4a5a-b6f7-df566b286601" containerName="nova-cell0-conductor-conductor" containerID="cri-o://9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec" gracePeriod=30 Dec 05 13:40:12 crc kubenswrapper[4809]: I1205 13:40:12.452543 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 13:40:12 crc kubenswrapper[4809]: I1205 13:40:12.452791 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="6644e4f7-66da-4ab0-8100-18ba3cac40bf" containerName="nova-cell1-conductor-conductor" containerID="cri-o://1ad11e73e4564fb1f0f0533d92ad600562ff60c4270a24cf1f1115b427f4097a" gracePeriod=30 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.150292 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.150817 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="49c8cdbe-f1e8-4d73-a86f-d18399642b85" containerName="nova-scheduler-scheduler" containerID="cri-o://db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658" gracePeriod=30 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.166779 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.167095 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-log" containerID="cri-o://4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec" gracePeriod=30 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.167175 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-api" containerID="cri-o://4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520" gracePeriod=30 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.181380 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.181663 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-log" containerID="cri-o://7330f21737c9df36e111348a792051200c138724d80a08c38a1d8dcbab55b678" gracePeriod=30 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.182219 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-metadata" containerID="cri-o://685061eb6291332a3626770b6c95c85cbdd8d9c9d41ad36969519a969ec3e998" gracePeriod=30 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.543213 4809 generic.go:334] "Generic (PLEG): container finished" podID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerID="7330f21737c9df36e111348a792051200c138724d80a08c38a1d8dcbab55b678" exitCode=143 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.543297 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51d9c44b-44e8-4bd6-876c-2047efc32605","Type":"ContainerDied","Data":"7330f21737c9df36e111348a792051200c138724d80a08c38a1d8dcbab55b678"} Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.550702 4809 generic.go:334] "Generic (PLEG): container finished" podID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerID="4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec" exitCode=143 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.550769 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b405ddbd-2a5f-4126-8c54-4bb55754c432","Type":"ContainerDied","Data":"4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec"} Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.553696 4809 generic.go:334] "Generic (PLEG): container finished" podID="6644e4f7-66da-4ab0-8100-18ba3cac40bf" containerID="1ad11e73e4564fb1f0f0533d92ad600562ff60c4270a24cf1f1115b427f4097a" exitCode=0 Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.553734 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6644e4f7-66da-4ab0-8100-18ba3cac40bf","Type":"ContainerDied","Data":"1ad11e73e4564fb1f0f0533d92ad600562ff60c4270a24cf1f1115b427f4097a"} Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.553757 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6644e4f7-66da-4ab0-8100-18ba3cac40bf","Type":"ContainerDied","Data":"3c34cd74a0a8fc04620d0abc4d63e8a74632cc2cacfc5316f9a2fa8e1d0c28f5"} Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.553772 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c34cd74a0a8fc04620d0abc4d63e8a74632cc2cacfc5316f9a2fa8e1d0c28f5" Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.553879 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.728639 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-config-data\") pod \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.728844 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgvvw\" (UniqueName: \"kubernetes.io/projected/6644e4f7-66da-4ab0-8100-18ba3cac40bf-kube-api-access-jgvvw\") pod \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.728883 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-combined-ca-bundle\") pod \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\" (UID: \"6644e4f7-66da-4ab0-8100-18ba3cac40bf\") " Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.733479 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6644e4f7-66da-4ab0-8100-18ba3cac40bf-kube-api-access-jgvvw" (OuterVolumeSpecName: "kube-api-access-jgvvw") pod "6644e4f7-66da-4ab0-8100-18ba3cac40bf" (UID: "6644e4f7-66da-4ab0-8100-18ba3cac40bf"). InnerVolumeSpecName "kube-api-access-jgvvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.766042 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-config-data" (OuterVolumeSpecName: "config-data") pod "6644e4f7-66da-4ab0-8100-18ba3cac40bf" (UID: "6644e4f7-66da-4ab0-8100-18ba3cac40bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.766078 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6644e4f7-66da-4ab0-8100-18ba3cac40bf" (UID: "6644e4f7-66da-4ab0-8100-18ba3cac40bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.831449 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgvvw\" (UniqueName: \"kubernetes.io/projected/6644e4f7-66da-4ab0-8100-18ba3cac40bf-kube-api-access-jgvvw\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.831846 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:13 crc kubenswrapper[4809]: I1205 13:40:13.831860 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6644e4f7-66da-4ab0-8100-18ba3cac40bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.563176 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.647373 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.673182 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.683774 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 13:40:14 crc kubenswrapper[4809]: E1205 13:40:14.684206 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6644e4f7-66da-4ab0-8100-18ba3cac40bf" containerName="nova-cell1-conductor-conductor" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684231 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="6644e4f7-66da-4ab0-8100-18ba3cac40bf" containerName="nova-cell1-conductor-conductor" Dec 05 13:40:14 crc kubenswrapper[4809]: E1205 13:40:14.684239 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3d10d8-24e2-4f12-9007-3fd1368261c0" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684247 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3d10d8-24e2-4f12-9007-3fd1368261c0" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 05 13:40:14 crc kubenswrapper[4809]: E1205 13:40:14.684278 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="extract-content" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684284 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="extract-content" Dec 05 13:40:14 crc kubenswrapper[4809]: E1205 13:40:14.684305 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="extract-utilities" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684312 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="extract-utilities" Dec 05 13:40:14 crc kubenswrapper[4809]: E1205 13:40:14.684326 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="registry-server" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684332 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="registry-server" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684549 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60bbc52-3283-4147-8ad8-28f9d0bb9281" containerName="registry-server" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684577 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="6644e4f7-66da-4ab0-8100-18ba3cac40bf" containerName="nova-cell1-conductor-conductor" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.684596 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b3d10d8-24e2-4f12-9007-3fd1368261c0" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.685371 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.694192 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.860477 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ae38166-059b-4dd8-b4df-939f676eeaac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.860783 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae38166-059b-4dd8-b4df-939f676eeaac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.860854 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xppjv\" (UniqueName: \"kubernetes.io/projected/0ae38166-059b-4dd8-b4df-939f676eeaac-kube-api-access-xppjv\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.887066 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6644e4f7-66da-4ab0-8100-18ba3cac40bf" path="/var/lib/kubelet/pods/6644e4f7-66da-4ab0-8100-18ba3cac40bf/volumes" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.963116 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ae38166-059b-4dd8-b4df-939f676eeaac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.963597 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae38166-059b-4dd8-b4df-939f676eeaac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.963663 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xppjv\" (UniqueName: \"kubernetes.io/projected/0ae38166-059b-4dd8-b4df-939f676eeaac-kube-api-access-xppjv\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.968532 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae38166-059b-4dd8-b4df-939f676eeaac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.969956 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ae38166-059b-4dd8-b4df-939f676eeaac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:14 crc kubenswrapper[4809]: I1205 13:40:14.990410 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xppjv\" (UniqueName: \"kubernetes.io/projected/0ae38166-059b-4dd8-b4df-939f676eeaac-kube-api-access-xppjv\") pod \"nova-cell1-conductor-0\" (UID: \"0ae38166-059b-4dd8-b4df-939f676eeaac\") " pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:15 crc kubenswrapper[4809]: I1205 13:40:15.019143 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:15 crc kubenswrapper[4809]: I1205 13:40:15.188296 4809 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 13:40:15 crc kubenswrapper[4809]: I1205 13:40:15.188365 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 13:40:15 crc kubenswrapper[4809]: I1205 13:40:15.681025 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 13:40:15 crc kubenswrapper[4809]: E1205 13:40:15.868461 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 13:40:15 crc kubenswrapper[4809]: E1205 13:40:15.871248 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 13:40:15 crc kubenswrapper[4809]: E1205 13:40:15.874447 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 13:40:15 crc kubenswrapper[4809]: E1205 13:40:15.874527 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="49c8cdbe-f1e8-4d73-a86f-d18399642b85" containerName="nova-scheduler-scheduler" Dec 05 13:40:16 crc kubenswrapper[4809]: E1205 13:40:16.021470 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 13:40:16 crc kubenswrapper[4809]: E1205 13:40:16.023048 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 13:40:16 crc kubenswrapper[4809]: E1205 13:40:16.025510 4809 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 05 13:40:16 crc kubenswrapper[4809]: E1205 13:40:16.025550 4809 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="daf4bee5-c20a-4a5a-b6f7-df566b286601" containerName="nova-cell0-conductor-conductor" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.151227 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 13:40:16 crc kubenswrapper[4809]: W1205 13:40:16.212678 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ae38166_059b_4dd8_b4df_939f676eeaac.slice/crio-cdc5da12188f7c23803fffbc0b91e22505d9d7eeaf07b0284b0fb309d8a78d46 WatchSource:0}: Error finding container cdc5da12188f7c23803fffbc0b91e22505d9d7eeaf07b0284b0fb309d8a78d46: Status 404 returned error can't find the container with id cdc5da12188f7c23803fffbc0b91e22505d9d7eeaf07b0284b0fb309d8a78d46 Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.594379 4809 generic.go:334] "Generic (PLEG): container finished" podID="daf4bee5-c20a-4a5a-b6f7-df566b286601" containerID="9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec" exitCode=0 Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.594462 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"daf4bee5-c20a-4a5a-b6f7-df566b286601","Type":"ContainerDied","Data":"9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec"} Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.601577 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.83:8775/\": read tcp 10.217.0.2:47522->10.217.1.83:8775: read: connection reset by peer" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.602045 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.83:8775/\": read tcp 10.217.0.2:47534->10.217.1.83:8775: read: connection reset by peer" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.602518 4809 generic.go:334] "Generic (PLEG): container finished" podID="49c8cdbe-f1e8-4d73-a86f-d18399642b85" containerID="db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658" exitCode=0 Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.602569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49c8cdbe-f1e8-4d73-a86f-d18399642b85","Type":"ContainerDied","Data":"db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658"} Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.613667 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0ae38166-059b-4dd8-b4df-939f676eeaac","Type":"ContainerStarted","Data":"4c7593c13967fb4364e46af0b2309e8fa555bc382b686a0141fcd8c5a98093de"} Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.613717 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0ae38166-059b-4dd8-b4df-939f676eeaac","Type":"ContainerStarted","Data":"cdc5da12188f7c23803fffbc0b91e22505d9d7eeaf07b0284b0fb309d8a78d46"} Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.614119 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.644403 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.644386893 podStartE2EDuration="2.644386893s" podCreationTimestamp="2025-12-05 13:40:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 13:40:16.641889946 +0000 UTC m=+9112.032866504" watchObservedRunningTime="2025-12-05 13:40:16.644386893 +0000 UTC m=+9112.035363451" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.844123 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 13:40:16 crc kubenswrapper[4809]: E1205 13:40:16.884204 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51d9c44b_44e8_4bd6_876c_2047efc32605.slice/crio-685061eb6291332a3626770b6c95c85cbdd8d9c9d41ad36969519a969ec3e998.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51d9c44b_44e8_4bd6_876c_2047efc32605.slice/crio-conmon-685061eb6291332a3626770b6c95c85cbdd8d9c9d41ad36969519a969ec3e998.scope\": RecentStats: unable to find data in memory cache]" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.891084 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.918811 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-combined-ca-bundle\") pod \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.918913 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwgjn\" (UniqueName: \"kubernetes.io/projected/daf4bee5-c20a-4a5a-b6f7-df566b286601-kube-api-access-qwgjn\") pod \"daf4bee5-c20a-4a5a-b6f7-df566b286601\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.918937 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-combined-ca-bundle\") pod \"daf4bee5-c20a-4a5a-b6f7-df566b286601\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.918968 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqtff\" (UniqueName: \"kubernetes.io/projected/49c8cdbe-f1e8-4d73-a86f-d18399642b85-kube-api-access-bqtff\") pod \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.919037 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-config-data\") pod \"daf4bee5-c20a-4a5a-b6f7-df566b286601\" (UID: \"daf4bee5-c20a-4a5a-b6f7-df566b286601\") " Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.919111 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-config-data\") pod \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\" (UID: \"49c8cdbe-f1e8-4d73-a86f-d18399642b85\") " Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.961836 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c8cdbe-f1e8-4d73-a86f-d18399642b85-kube-api-access-bqtff" (OuterVolumeSpecName: "kube-api-access-bqtff") pod "49c8cdbe-f1e8-4d73-a86f-d18399642b85" (UID: "49c8cdbe-f1e8-4d73-a86f-d18399642b85"). InnerVolumeSpecName "kube-api-access-bqtff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.964409 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf4bee5-c20a-4a5a-b6f7-df566b286601-kube-api-access-qwgjn" (OuterVolumeSpecName: "kube-api-access-qwgjn") pod "daf4bee5-c20a-4a5a-b6f7-df566b286601" (UID: "daf4bee5-c20a-4a5a-b6f7-df566b286601"). InnerVolumeSpecName "kube-api-access-qwgjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.977090 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-config-data" (OuterVolumeSpecName: "config-data") pod "daf4bee5-c20a-4a5a-b6f7-df566b286601" (UID: "daf4bee5-c20a-4a5a-b6f7-df566b286601"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.978562 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49c8cdbe-f1e8-4d73-a86f-d18399642b85" (UID: "49c8cdbe-f1e8-4d73-a86f-d18399642b85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.985729 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-config-data" (OuterVolumeSpecName: "config-data") pod "49c8cdbe-f1e8-4d73-a86f-d18399642b85" (UID: "49c8cdbe-f1e8-4d73-a86f-d18399642b85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:16 crc kubenswrapper[4809]: I1205 13:40:16.992966 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "daf4bee5-c20a-4a5a-b6f7-df566b286601" (UID: "daf4bee5-c20a-4a5a-b6f7-df566b286601"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.009034 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.020915 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-combined-ca-bundle\") pod \"b405ddbd-2a5f-4126-8c54-4bb55754c432\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.020999 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d5dz\" (UniqueName: \"kubernetes.io/projected/b405ddbd-2a5f-4126-8c54-4bb55754c432-kube-api-access-2d5dz\") pod \"b405ddbd-2a5f-4126-8c54-4bb55754c432\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.021092 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-config-data\") pod \"b405ddbd-2a5f-4126-8c54-4bb55754c432\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.021218 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b405ddbd-2a5f-4126-8c54-4bb55754c432-logs\") pod \"b405ddbd-2a5f-4126-8c54-4bb55754c432\" (UID: \"b405ddbd-2a5f-4126-8c54-4bb55754c432\") " Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.021998 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.022016 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwgjn\" (UniqueName: \"kubernetes.io/projected/daf4bee5-c20a-4a5a-b6f7-df566b286601-kube-api-access-qwgjn\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.022027 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.022037 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqtff\" (UniqueName: \"kubernetes.io/projected/49c8cdbe-f1e8-4d73-a86f-d18399642b85-kube-api-access-bqtff\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.022045 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf4bee5-c20a-4a5a-b6f7-df566b286601-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.022055 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49c8cdbe-f1e8-4d73-a86f-d18399642b85-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.022164 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b405ddbd-2a5f-4126-8c54-4bb55754c432-logs" (OuterVolumeSpecName: "logs") pod "b405ddbd-2a5f-4126-8c54-4bb55754c432" (UID: "b405ddbd-2a5f-4126-8c54-4bb55754c432"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.035141 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b405ddbd-2a5f-4126-8c54-4bb55754c432-kube-api-access-2d5dz" (OuterVolumeSpecName: "kube-api-access-2d5dz") pod "b405ddbd-2a5f-4126-8c54-4bb55754c432" (UID: "b405ddbd-2a5f-4126-8c54-4bb55754c432"). InnerVolumeSpecName "kube-api-access-2d5dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.077181 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b405ddbd-2a5f-4126-8c54-4bb55754c432" (UID: "b405ddbd-2a5f-4126-8c54-4bb55754c432"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.078422 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-config-data" (OuterVolumeSpecName: "config-data") pod "b405ddbd-2a5f-4126-8c54-4bb55754c432" (UID: "b405ddbd-2a5f-4126-8c54-4bb55754c432"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.124205 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.124258 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b405ddbd-2a5f-4126-8c54-4bb55754c432-logs\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.124269 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b405ddbd-2a5f-4126-8c54-4bb55754c432-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.124285 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d5dz\" (UniqueName: \"kubernetes.io/projected/b405ddbd-2a5f-4126-8c54-4bb55754c432-kube-api-access-2d5dz\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.583712 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv"] Dec 05 13:40:17 crc kubenswrapper[4809]: E1205 13:40:17.584294 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf4bee5-c20a-4a5a-b6f7-df566b286601" containerName="nova-cell0-conductor-conductor" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584315 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf4bee5-c20a-4a5a-b6f7-df566b286601" containerName="nova-cell0-conductor-conductor" Dec 05 13:40:17 crc kubenswrapper[4809]: E1205 13:40:17.584326 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-api" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584334 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-api" Dec 05 13:40:17 crc kubenswrapper[4809]: E1205 13:40:17.584345 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-log" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584351 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-log" Dec 05 13:40:17 crc kubenswrapper[4809]: E1205 13:40:17.584390 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c8cdbe-f1e8-4d73-a86f-d18399642b85" containerName="nova-scheduler-scheduler" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584397 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c8cdbe-f1e8-4d73-a86f-d18399642b85" containerName="nova-scheduler-scheduler" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584614 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-log" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584651 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="49c8cdbe-f1e8-4d73-a86f-d18399642b85" containerName="nova-scheduler-scheduler" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584661 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf4bee5-c20a-4a5a-b6f7-df566b286601" containerName="nova-cell0-conductor-conductor" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.584688 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerName="nova-api-api" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.586596 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.589310 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.589552 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.589896 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.589949 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.590237 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-fxqbq" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.590368 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.590665 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.630950 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv"] Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.667457 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.668299 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"daf4bee5-c20a-4a5a-b6f7-df566b286601","Type":"ContainerDied","Data":"a1e6c813fa8c2a540a5bd9ef067cf6bb8db7dd5f9d11893185a68459b76b0fed"} Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.668354 4809 scope.go:117] "RemoveContainer" containerID="9cb1f39d14acf99e578cb62ebf8f8ec2b53f3c0977caad3c5ce1b8db298ffcec" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.682350 4809 generic.go:334] "Generic (PLEG): container finished" podID="b405ddbd-2a5f-4126-8c54-4bb55754c432" containerID="4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520" exitCode=0 Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.682470 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b405ddbd-2a5f-4126-8c54-4bb55754c432","Type":"ContainerDied","Data":"4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520"} Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.682502 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b405ddbd-2a5f-4126-8c54-4bb55754c432","Type":"ContainerDied","Data":"628e49c1023e0e868b0f8979616333eee6251bc8cc72f73847542f0a40d5d798"} Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.682682 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.698577 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"49c8cdbe-f1e8-4d73-a86f-d18399642b85","Type":"ContainerDied","Data":"84155bcfb21d37c0ad797cc4a9373c633d4dc8a0a903408018d210af5481143f"} Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.698763 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.704894 4809 generic.go:334] "Generic (PLEG): container finished" podID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerID="685061eb6291332a3626770b6c95c85cbdd8d9c9d41ad36969519a969ec3e998" exitCode=0 Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.704978 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51d9c44b-44e8-4bd6-876c-2047efc32605","Type":"ContainerDied","Data":"685061eb6291332a3626770b6c95c85cbdd8d9c9d41ad36969519a969ec3e998"} Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.749843 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750312 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750348 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750467 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750497 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750524 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750585 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750666 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nqp7\" (UniqueName: \"kubernetes.io/projected/2959aa37-7f80-4b96-9284-c508595cce86-kube-api-access-4nqp7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.750713 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.751279 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.751335 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853167 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853260 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853410 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853434 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853553 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853577 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853597 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853643 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853666 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nqp7\" (UniqueName: \"kubernetes.io/projected/2959aa37-7f80-4b96-9284-c508595cce86-kube-api-access-4nqp7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853717 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.853801 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.854978 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.855800 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.857298 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.857784 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.858008 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.858442 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.859530 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.860157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.861173 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.869109 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:17 crc kubenswrapper[4809]: I1205 13:40:17.981491 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nqp7\" (UniqueName: \"kubernetes.io/projected/2959aa37-7f80-4b96-9284-c508595cce86-kube-api-access-4nqp7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.011502 4809 scope.go:117] "RemoveContainer" containerID="4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.066028 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.068271 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.070351 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.091122 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.094403 4809 scope.go:117] "RemoveContainer" containerID="4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.112406 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.137568 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.151779 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: E1205 13:40:18.152615 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-log" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.152686 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-log" Dec 05 13:40:18 crc kubenswrapper[4809]: E1205 13:40:18.152717 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-metadata" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.152726 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-metadata" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.152996 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-metadata" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.153019 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" containerName="nova-metadata-log" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.153896 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.158171 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.162740 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.170032 4809 scope.go:117] "RemoveContainer" containerID="4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520" Dec 05 13:40:18 crc kubenswrapper[4809]: E1205 13:40:18.170530 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520\": container with ID starting with 4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520 not found: ID does not exist" containerID="4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.170572 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520"} err="failed to get container status \"4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520\": rpc error: code = NotFound desc = could not find container \"4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520\": container with ID starting with 4d812091b8815935b98955d5de9a7867bdd429fb356034f870038dcc0e56e520 not found: ID does not exist" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.170598 4809 scope.go:117] "RemoveContainer" containerID="4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec" Dec 05 13:40:18 crc kubenswrapper[4809]: E1205 13:40:18.170901 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec\": container with ID starting with 4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec not found: ID does not exist" containerID="4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.170997 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec"} err="failed to get container status \"4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec\": rpc error: code = NotFound desc = could not find container \"4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec\": container with ID starting with 4b9137d614af0fa09bc6541467ac04f417a5e90cb06ba8ef7c39fbcf0e5300ec not found: ID does not exist" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.171078 4809 scope.go:117] "RemoveContainer" containerID="db181798403a4dfdf12c0a7373527751cf4099be476af57a9e07090b19b56658" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.171579 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51d9c44b-44e8-4bd6-876c-2047efc32605-logs\") pod \"51d9c44b-44e8-4bd6-876c-2047efc32605\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.171646 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf8cv\" (UniqueName: \"kubernetes.io/projected/51d9c44b-44e8-4bd6-876c-2047efc32605-kube-api-access-rf8cv\") pod \"51d9c44b-44e8-4bd6-876c-2047efc32605\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.171689 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-combined-ca-bundle\") pod \"51d9c44b-44e8-4bd6-876c-2047efc32605\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.171719 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-config-data\") pod \"51d9c44b-44e8-4bd6-876c-2047efc32605\" (UID: \"51d9c44b-44e8-4bd6-876c-2047efc32605\") " Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.172370 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e73347-b4a0-43ef-9a9c-022a54af26b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.172405 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd76d\" (UniqueName: \"kubernetes.io/projected/b0e73347-b4a0-43ef-9a9c-022a54af26b7-kube-api-access-dd76d\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.172510 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e73347-b4a0-43ef-9a9c-022a54af26b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.179442 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51d9c44b-44e8-4bd6-876c-2047efc32605-logs" (OuterVolumeSpecName: "logs") pod "51d9c44b-44e8-4bd6-876c-2047efc32605" (UID: "51d9c44b-44e8-4bd6-876c-2047efc32605"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.182088 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.200550 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51d9c44b-44e8-4bd6-876c-2047efc32605-kube-api-access-rf8cv" (OuterVolumeSpecName: "kube-api-access-rf8cv") pod "51d9c44b-44e8-4bd6-876c-2047efc32605" (UID: "51d9c44b-44e8-4bd6-876c-2047efc32605"). InnerVolumeSpecName "kube-api-access-rf8cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.207867 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.211691 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.214442 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.224938 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.228149 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-config-data" (OuterVolumeSpecName: "config-data") pod "51d9c44b-44e8-4bd6-876c-2047efc32605" (UID: "51d9c44b-44e8-4bd6-876c-2047efc32605"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.233546 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51d9c44b-44e8-4bd6-876c-2047efc32605" (UID: "51d9c44b-44e8-4bd6-876c-2047efc32605"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.253969 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.258176 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.260541 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274478 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3544290a-8953-442b-85af-1ae538e6115c-config-data\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274564 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e73347-b4a0-43ef-9a9c-022a54af26b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274599 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd76d\" (UniqueName: \"kubernetes.io/projected/b0e73347-b4a0-43ef-9a9c-022a54af26b7-kube-api-access-dd76d\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274671 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d29gf\" (UniqueName: \"kubernetes.io/projected/3544290a-8953-442b-85af-1ae538e6115c-kube-api-access-d29gf\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274709 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e73347-b4a0-43ef-9a9c-022a54af26b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274840 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjmcj\" (UniqueName: \"kubernetes.io/projected/edda60ef-c084-4473-9ca8-c5f755a3248b-kube-api-access-rjmcj\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274883 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edda60ef-c084-4473-9ca8-c5f755a3248b-logs\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274928 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edda60ef-c084-4473-9ca8-c5f755a3248b-config-data\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.274963 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edda60ef-c084-4473-9ca8-c5f755a3248b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.275003 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3544290a-8953-442b-85af-1ae538e6115c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.275122 4809 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51d9c44b-44e8-4bd6-876c-2047efc32605-logs\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.275136 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf8cv\" (UniqueName: \"kubernetes.io/projected/51d9c44b-44e8-4bd6-876c-2047efc32605-kube-api-access-rf8cv\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.275149 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.275160 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d9c44b-44e8-4bd6-876c-2047efc32605-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.279283 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.282984 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e73347-b4a0-43ef-9a9c-022a54af26b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.283897 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e73347-b4a0-43ef-9a9c-022a54af26b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.291131 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.300445 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd76d\" (UniqueName: \"kubernetes.io/projected/b0e73347-b4a0-43ef-9a9c-022a54af26b7-kube-api-access-dd76d\") pod \"nova-cell0-conductor-0\" (UID: \"b0e73347-b4a0-43ef-9a9c-022a54af26b7\") " pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.377406 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3544290a-8953-442b-85af-1ae538e6115c-config-data\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.377756 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d29gf\" (UniqueName: \"kubernetes.io/projected/3544290a-8953-442b-85af-1ae538e6115c-kube-api-access-d29gf\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.377862 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjmcj\" (UniqueName: \"kubernetes.io/projected/edda60ef-c084-4473-9ca8-c5f755a3248b-kube-api-access-rjmcj\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.377900 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edda60ef-c084-4473-9ca8-c5f755a3248b-logs\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.377941 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edda60ef-c084-4473-9ca8-c5f755a3248b-config-data\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.377972 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edda60ef-c084-4473-9ca8-c5f755a3248b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.377999 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3544290a-8953-442b-85af-1ae538e6115c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.381972 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3544290a-8953-442b-85af-1ae538e6115c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.382252 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edda60ef-c084-4473-9ca8-c5f755a3248b-logs\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.386621 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edda60ef-c084-4473-9ca8-c5f755a3248b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.386952 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edda60ef-c084-4473-9ca8-c5f755a3248b-config-data\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.402423 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjmcj\" (UniqueName: \"kubernetes.io/projected/edda60ef-c084-4473-9ca8-c5f755a3248b-kube-api-access-rjmcj\") pod \"nova-api-0\" (UID: \"edda60ef-c084-4473-9ca8-c5f755a3248b\") " pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.403280 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d29gf\" (UniqueName: \"kubernetes.io/projected/3544290a-8953-442b-85af-1ae538e6115c-kube-api-access-d29gf\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.411473 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3544290a-8953-442b-85af-1ae538e6115c-config-data\") pod \"nova-scheduler-0\" (UID: \"3544290a-8953-442b-85af-1ae538e6115c\") " pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.473362 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.487292 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.680425 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.728581 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"51d9c44b-44e8-4bd6-876c-2047efc32605","Type":"ContainerDied","Data":"4b756ba9cea93acd2ba9b7d66bd8e2ef9597fab25831b824ba004cb8f6f82ea3"} Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.728621 4809 scope.go:117] "RemoveContainer" containerID="685061eb6291332a3626770b6c95c85cbdd8d9c9d41ad36969519a969ec3e998" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.728751 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.794029 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.815343 4809 scope.go:117] "RemoveContainer" containerID="7330f21737c9df36e111348a792051200c138724d80a08c38a1d8dcbab55b678" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.823963 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.846436 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.848892 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.851965 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.891255 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003a4f36-e5aa-482f-af8e-bd1297698e12-config-data\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.891346 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003a4f36-e5aa-482f-af8e-bd1297698e12-logs\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.891385 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003a4f36-e5aa-482f-af8e-bd1297698e12-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.891478 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fp59\" (UniqueName: \"kubernetes.io/projected/003a4f36-e5aa-482f-af8e-bd1297698e12-kube-api-access-9fp59\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: W1205 13:40:18.898746 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2959aa37_7f80_4b96_9284_c508595cce86.slice/crio-f19804925f034716756433db8dd34b19f1ca345389f9aded2aa49bafa64495ad WatchSource:0}: Error finding container f19804925f034716756433db8dd34b19f1ca345389f9aded2aa49bafa64495ad: Status 404 returned error can't find the container with id f19804925f034716756433db8dd34b19f1ca345389f9aded2aa49bafa64495ad Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.908468 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c8cdbe-f1e8-4d73-a86f-d18399642b85" path="/var/lib/kubelet/pods/49c8cdbe-f1e8-4d73-a86f-d18399642b85/volumes" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.909348 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51d9c44b-44e8-4bd6-876c-2047efc32605" path="/var/lib/kubelet/pods/51d9c44b-44e8-4bd6-876c-2047efc32605/volumes" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.910359 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b405ddbd-2a5f-4126-8c54-4bb55754c432" path="/var/lib/kubelet/pods/b405ddbd-2a5f-4126-8c54-4bb55754c432/volumes" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.911880 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf4bee5-c20a-4a5a-b6f7-df566b286601" path="/var/lib/kubelet/pods/daf4bee5-c20a-4a5a-b6f7-df566b286601/volumes" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.912560 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.912595 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv"] Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.993369 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fp59\" (UniqueName: \"kubernetes.io/projected/003a4f36-e5aa-482f-af8e-bd1297698e12-kube-api-access-9fp59\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.993558 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003a4f36-e5aa-482f-af8e-bd1297698e12-config-data\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.993657 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003a4f36-e5aa-482f-af8e-bd1297698e12-logs\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.993695 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003a4f36-e5aa-482f-af8e-bd1297698e12-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:18 crc kubenswrapper[4809]: I1205 13:40:18.994225 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/003a4f36-e5aa-482f-af8e-bd1297698e12-logs\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.000598 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003a4f36-e5aa-482f-af8e-bd1297698e12-config-data\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.000743 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003a4f36-e5aa-482f-af8e-bd1297698e12-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.017709 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fp59\" (UniqueName: \"kubernetes.io/projected/003a4f36-e5aa-482f-af8e-bd1297698e12-kube-api-access-9fp59\") pod \"nova-metadata-0\" (UID: \"003a4f36-e5aa-482f-af8e-bd1297698e12\") " pod="openstack/nova-metadata-0" Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.078797 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.091715 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.180911 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.364176 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 13:40:19 crc kubenswrapper[4809]: W1205 13:40:19.437741 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedda60ef_c084_4473_9ca8_c5f755a3248b.slice/crio-1c4832bb599b1d411421e64a3a4a9d04fb0c398ac8a936ce3f5609e2598ee41f WatchSource:0}: Error finding container 1c4832bb599b1d411421e64a3a4a9d04fb0c398ac8a936ce3f5609e2598ee41f: Status 404 returned error can't find the container with id 1c4832bb599b1d411421e64a3a4a9d04fb0c398ac8a936ce3f5609e2598ee41f Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.739912 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3544290a-8953-442b-85af-1ae538e6115c","Type":"ContainerStarted","Data":"941482d7499587da25dd03ef5b66c2ecc9d212a56c471743e9c1a0a7f606aea7"} Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.741251 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b0e73347-b4a0-43ef-9a9c-022a54af26b7","Type":"ContainerStarted","Data":"2b91117a6d5f147158b6e5a52e1010b9680dc868da225ba7590fb870fc88bb80"} Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.742326 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" event={"ID":"2959aa37-7f80-4b96-9284-c508595cce86","Type":"ContainerStarted","Data":"f19804925f034716756433db8dd34b19f1ca345389f9aded2aa49bafa64495ad"} Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.743328 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"edda60ef-c084-4473-9ca8-c5f755a3248b","Type":"ContainerStarted","Data":"1c4832bb599b1d411421e64a3a4a9d04fb0c398ac8a936ce3f5609e2598ee41f"} Dec 05 13:40:19 crc kubenswrapper[4809]: I1205 13:40:19.897738 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 13:40:19 crc kubenswrapper[4809]: W1205 13:40:19.899415 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod003a4f36_e5aa_482f_af8e_bd1297698e12.slice/crio-701f9de65c0a8be63e56ee45e45860901064c304bab3bb56d0ec4110d7272ed4 WatchSource:0}: Error finding container 701f9de65c0a8be63e56ee45e45860901064c304bab3bb56d0ec4110d7272ed4: Status 404 returned error can't find the container with id 701f9de65c0a8be63e56ee45e45860901064c304bab3bb56d0ec4110d7272ed4 Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.762087 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"003a4f36-e5aa-482f-af8e-bd1297698e12","Type":"ContainerStarted","Data":"1e8d0fda3128c3b107682e405e44f5d1251fabae3d7e0ba1432725acc8f298fe"} Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.762429 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"003a4f36-e5aa-482f-af8e-bd1297698e12","Type":"ContainerStarted","Data":"701f9de65c0a8be63e56ee45e45860901064c304bab3bb56d0ec4110d7272ed4"} Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.764376 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3544290a-8953-442b-85af-1ae538e6115c","Type":"ContainerStarted","Data":"c31c4ca7aaa710c7c34bcbc14d78637052c76285241bb38f528025543ab9cfd3"} Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.768364 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b0e73347-b4a0-43ef-9a9c-022a54af26b7","Type":"ContainerStarted","Data":"027d4e2b28002be14e0e978007eba490a79ea4bf7fd7a4951dede7f2326a4815"} Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.768591 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.771047 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"edda60ef-c084-4473-9ca8-c5f755a3248b","Type":"ContainerStarted","Data":"5722de69b60bbd63e207ce17d6a51302f188914044f7c69421ef2c4e9d74e073"} Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.793621 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.793596269 podStartE2EDuration="2.793596269s" podCreationTimestamp="2025-12-05 13:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 13:40:20.783707995 +0000 UTC m=+9116.174684573" watchObservedRunningTime="2025-12-05 13:40:20.793596269 +0000 UTC m=+9116.184572827" Dec 05 13:40:20 crc kubenswrapper[4809]: I1205 13:40:20.810688 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.810612563 podStartE2EDuration="2.810612563s" podCreationTimestamp="2025-12-05 13:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 13:40:20.79812299 +0000 UTC m=+9116.189099568" watchObservedRunningTime="2025-12-05 13:40:20.810612563 +0000 UTC m=+9116.201589121" Dec 05 13:40:21 crc kubenswrapper[4809]: I1205 13:40:21.786520 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" event={"ID":"2959aa37-7f80-4b96-9284-c508595cce86","Type":"ContainerStarted","Data":"0558eee863b24d0be7642f71c019a3f9da7c11ce41aeb81e15f189c59e35dd0b"} Dec 05 13:40:21 crc kubenswrapper[4809]: I1205 13:40:21.788825 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"edda60ef-c084-4473-9ca8-c5f755a3248b","Type":"ContainerStarted","Data":"c26d4b5e3a0ee487bcdc2555d37c2bb7468891fb6fc1802250a032ff68bd6ed4"} Dec 05 13:40:21 crc kubenswrapper[4809]: I1205 13:40:21.791383 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"003a4f36-e5aa-482f-af8e-bd1297698e12","Type":"ContainerStarted","Data":"870f5d0d6568a4dffda28d2eb4d03808498216d955426673f0b768adfa3e5e34"} Dec 05 13:40:21 crc kubenswrapper[4809]: I1205 13:40:21.811657 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" podStartSLOduration=3.190104473 podStartE2EDuration="4.811640199s" podCreationTimestamp="2025-12-05 13:40:17 +0000 UTC" firstStartedPulling="2025-12-05 13:40:18.904957494 +0000 UTC m=+9114.295934052" lastFinishedPulling="2025-12-05 13:40:20.52649322 +0000 UTC m=+9115.917469778" observedRunningTime="2025-12-05 13:40:21.809724958 +0000 UTC m=+9117.200701516" watchObservedRunningTime="2025-12-05 13:40:21.811640199 +0000 UTC m=+9117.202616747" Dec 05 13:40:21 crc kubenswrapper[4809]: I1205 13:40:21.857162 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.857139733 podStartE2EDuration="3.857139733s" podCreationTimestamp="2025-12-05 13:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 13:40:21.856325932 +0000 UTC m=+9117.247302500" watchObservedRunningTime="2025-12-05 13:40:21.857139733 +0000 UTC m=+9117.248116291" Dec 05 13:40:21 crc kubenswrapper[4809]: I1205 13:40:21.865236 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.865216459 podStartE2EDuration="3.865216459s" podCreationTimestamp="2025-12-05 13:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 13:40:21.836117852 +0000 UTC m=+9117.227094420" watchObservedRunningTime="2025-12-05 13:40:21.865216459 +0000 UTC m=+9117.256193017" Dec 05 13:40:23 crc kubenswrapper[4809]: I1205 13:40:23.487892 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 13:40:24 crc kubenswrapper[4809]: I1205 13:40:24.181790 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 13:40:24 crc kubenswrapper[4809]: I1205 13:40:24.181837 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 13:40:25 crc kubenswrapper[4809]: I1205 13:40:25.051620 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 13:40:28 crc kubenswrapper[4809]: I1205 13:40:28.487832 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 13:40:28 crc kubenswrapper[4809]: I1205 13:40:28.523461 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 13:40:28 crc kubenswrapper[4809]: I1205 13:40:28.523786 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 13:40:28 crc kubenswrapper[4809]: I1205 13:40:28.682213 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 13:40:28 crc kubenswrapper[4809]: I1205 13:40:28.682552 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 13:40:28 crc kubenswrapper[4809]: I1205 13:40:28.898755 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 13:40:29 crc kubenswrapper[4809]: I1205 13:40:29.182057 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 13:40:29 crc kubenswrapper[4809]: I1205 13:40:29.182134 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 13:40:29 crc kubenswrapper[4809]: I1205 13:40:29.764933 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="edda60ef-c084-4473-9ca8-c5f755a3248b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 13:40:29 crc kubenswrapper[4809]: I1205 13:40:29.764936 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="edda60ef-c084-4473-9ca8-c5f755a3248b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 13:40:30 crc kubenswrapper[4809]: I1205 13:40:30.264903 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="003a4f36-e5aa-482f-af8e-bd1297698e12" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.198:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 13:40:30 crc kubenswrapper[4809]: I1205 13:40:30.265539 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="003a4f36-e5aa-482f-af8e-bd1297698e12" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.198:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 13:40:38 crc kubenswrapper[4809]: I1205 13:40:38.687012 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 13:40:38 crc kubenswrapper[4809]: I1205 13:40:38.688109 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 13:40:38 crc kubenswrapper[4809]: I1205 13:40:38.688338 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 13:40:38 crc kubenswrapper[4809]: I1205 13:40:38.694100 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 13:40:38 crc kubenswrapper[4809]: I1205 13:40:38.994360 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 13:40:38 crc kubenswrapper[4809]: I1205 13:40:38.997426 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 13:40:39 crc kubenswrapper[4809]: I1205 13:40:39.197095 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 13:40:39 crc kubenswrapper[4809]: I1205 13:40:39.201274 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 13:40:39 crc kubenswrapper[4809]: I1205 13:40:39.203180 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 13:40:40 crc kubenswrapper[4809]: I1205 13:40:40.022986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 13:40:44 crc kubenswrapper[4809]: I1205 13:40:44.047039 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:40:44 crc kubenswrapper[4809]: I1205 13:40:44.047515 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:40:47 crc kubenswrapper[4809]: I1205 13:40:47.403167 4809 scope.go:117] "RemoveContainer" containerID="1ad11e73e4564fb1f0f0533d92ad600562ff60c4270a24cf1f1115b427f4097a" Dec 05 13:41:14 crc kubenswrapper[4809]: I1205 13:41:14.046398 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:41:14 crc kubenswrapper[4809]: I1205 13:41:14.047038 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.047001 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.047516 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.047561 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.048288 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.048355 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" gracePeriod=600 Dec 05 13:41:44 crc kubenswrapper[4809]: E1205 13:41:44.169375 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.783052 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" exitCode=0 Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.783127 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103"} Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.783382 4809 scope.go:117] "RemoveContainer" containerID="ee2b1b16721c3b6a23331471a609811f8cfbcb2367cda45319070a7cccde75a2" Dec 05 13:41:44 crc kubenswrapper[4809]: I1205 13:41:44.784096 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:41:44 crc kubenswrapper[4809]: E1205 13:41:44.784423 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:41:56 crc kubenswrapper[4809]: I1205 13:41:56.873244 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:41:56 crc kubenswrapper[4809]: E1205 13:41:56.874026 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:42:07 crc kubenswrapper[4809]: I1205 13:42:07.873409 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:42:07 crc kubenswrapper[4809]: E1205 13:42:07.874694 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:42:12 crc kubenswrapper[4809]: I1205 13:42:12.935746 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-84wf8"] Dec 05 13:42:12 crc kubenswrapper[4809]: I1205 13:42:12.939874 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:12 crc kubenswrapper[4809]: I1205 13:42:12.952297 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84wf8"] Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.003910 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqdmp\" (UniqueName: \"kubernetes.io/projected/b299a302-8020-4c2d-b9da-72d23d9dc042-kube-api-access-hqdmp\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.004026 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-catalog-content\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.004140 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-utilities\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.106570 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-utilities\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.106760 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqdmp\" (UniqueName: \"kubernetes.io/projected/b299a302-8020-4c2d-b9da-72d23d9dc042-kube-api-access-hqdmp\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.107251 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-utilities\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.107552 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-catalog-content\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.108006 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-catalog-content\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.132497 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqdmp\" (UniqueName: \"kubernetes.io/projected/b299a302-8020-4c2d-b9da-72d23d9dc042-kube-api-access-hqdmp\") pod \"certified-operators-84wf8\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:13 crc kubenswrapper[4809]: I1205 13:42:13.279189 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:14 crc kubenswrapper[4809]: I1205 13:42:14.252274 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84wf8"] Dec 05 13:42:15 crc kubenswrapper[4809]: I1205 13:42:15.150577 4809 generic.go:334] "Generic (PLEG): container finished" podID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerID="66fef099b727323e611796072bc21363682d5c573f8ab466172a7593adb0c001" exitCode=0 Dec 05 13:42:15 crc kubenswrapper[4809]: I1205 13:42:15.150639 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84wf8" event={"ID":"b299a302-8020-4c2d-b9da-72d23d9dc042","Type":"ContainerDied","Data":"66fef099b727323e611796072bc21363682d5c573f8ab466172a7593adb0c001"} Dec 05 13:42:15 crc kubenswrapper[4809]: I1205 13:42:15.151139 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84wf8" event={"ID":"b299a302-8020-4c2d-b9da-72d23d9dc042","Type":"ContainerStarted","Data":"d9e0b36154072999ccbd6455180eb29824036884537f7718580bbf37a1c237cc"} Dec 05 13:42:17 crc kubenswrapper[4809]: I1205 13:42:17.209324 4809 generic.go:334] "Generic (PLEG): container finished" podID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerID="cfb131b2bbb5829e3613763865ebffc1eaed49117d67d975c8c06c199af0464c" exitCode=0 Dec 05 13:42:17 crc kubenswrapper[4809]: I1205 13:42:17.209453 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84wf8" event={"ID":"b299a302-8020-4c2d-b9da-72d23d9dc042","Type":"ContainerDied","Data":"cfb131b2bbb5829e3613763865ebffc1eaed49117d67d975c8c06c199af0464c"} Dec 05 13:42:18 crc kubenswrapper[4809]: I1205 13:42:18.225243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84wf8" event={"ID":"b299a302-8020-4c2d-b9da-72d23d9dc042","Type":"ContainerStarted","Data":"2e59ffb0b75f8f345aba10855bad516d85214ed2596c9560139a3472b0c96511"} Dec 05 13:42:18 crc kubenswrapper[4809]: I1205 13:42:18.253844 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-84wf8" podStartSLOduration=3.69155489 podStartE2EDuration="6.253822874s" podCreationTimestamp="2025-12-05 13:42:12 +0000 UTC" firstStartedPulling="2025-12-05 13:42:15.152918844 +0000 UTC m=+9230.543895402" lastFinishedPulling="2025-12-05 13:42:17.715186828 +0000 UTC m=+9233.106163386" observedRunningTime="2025-12-05 13:42:18.244816523 +0000 UTC m=+9233.635793081" watchObservedRunningTime="2025-12-05 13:42:18.253822874 +0000 UTC m=+9233.644799422" Dec 05 13:42:18 crc kubenswrapper[4809]: I1205 13:42:18.872620 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:42:18 crc kubenswrapper[4809]: E1205 13:42:18.873234 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:42:23 crc kubenswrapper[4809]: I1205 13:42:23.279892 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:23 crc kubenswrapper[4809]: I1205 13:42:23.280624 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:23 crc kubenswrapper[4809]: I1205 13:42:23.346384 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:24 crc kubenswrapper[4809]: I1205 13:42:24.358948 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:24 crc kubenswrapper[4809]: I1205 13:42:24.417858 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84wf8"] Dec 05 13:42:26 crc kubenswrapper[4809]: I1205 13:42:26.317251 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-84wf8" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="registry-server" containerID="cri-o://2e59ffb0b75f8f345aba10855bad516d85214ed2596c9560139a3472b0c96511" gracePeriod=2 Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.330073 4809 generic.go:334] "Generic (PLEG): container finished" podID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerID="2e59ffb0b75f8f345aba10855bad516d85214ed2596c9560139a3472b0c96511" exitCode=0 Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.330144 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84wf8" event={"ID":"b299a302-8020-4c2d-b9da-72d23d9dc042","Type":"ContainerDied","Data":"2e59ffb0b75f8f345aba10855bad516d85214ed2596c9560139a3472b0c96511"} Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.586640 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.725623 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-catalog-content\") pod \"b299a302-8020-4c2d-b9da-72d23d9dc042\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.726067 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-utilities\") pod \"b299a302-8020-4c2d-b9da-72d23d9dc042\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.726123 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqdmp\" (UniqueName: \"kubernetes.io/projected/b299a302-8020-4c2d-b9da-72d23d9dc042-kube-api-access-hqdmp\") pod \"b299a302-8020-4c2d-b9da-72d23d9dc042\" (UID: \"b299a302-8020-4c2d-b9da-72d23d9dc042\") " Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.727287 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-utilities" (OuterVolumeSpecName: "utilities") pod "b299a302-8020-4c2d-b9da-72d23d9dc042" (UID: "b299a302-8020-4c2d-b9da-72d23d9dc042"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.732565 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b299a302-8020-4c2d-b9da-72d23d9dc042-kube-api-access-hqdmp" (OuterVolumeSpecName: "kube-api-access-hqdmp") pod "b299a302-8020-4c2d-b9da-72d23d9dc042" (UID: "b299a302-8020-4c2d-b9da-72d23d9dc042"). InnerVolumeSpecName "kube-api-access-hqdmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.828296 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:42:27 crc kubenswrapper[4809]: I1205 13:42:27.828558 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqdmp\" (UniqueName: \"kubernetes.io/projected/b299a302-8020-4c2d-b9da-72d23d9dc042-kube-api-access-hqdmp\") on node \"crc\" DevicePath \"\"" Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.218010 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b299a302-8020-4c2d-b9da-72d23d9dc042" (UID: "b299a302-8020-4c2d-b9da-72d23d9dc042"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.245010 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b299a302-8020-4c2d-b9da-72d23d9dc042-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.350612 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84wf8" event={"ID":"b299a302-8020-4c2d-b9da-72d23d9dc042","Type":"ContainerDied","Data":"d9e0b36154072999ccbd6455180eb29824036884537f7718580bbf37a1c237cc"} Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.350672 4809 scope.go:117] "RemoveContainer" containerID="2e59ffb0b75f8f345aba10855bad516d85214ed2596c9560139a3472b0c96511" Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.350685 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84wf8" Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.379194 4809 scope.go:117] "RemoveContainer" containerID="cfb131b2bbb5829e3613763865ebffc1eaed49117d67d975c8c06c199af0464c" Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.418927 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84wf8"] Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.427950 4809 scope.go:117] "RemoveContainer" containerID="66fef099b727323e611796072bc21363682d5c573f8ab466172a7593adb0c001" Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.436147 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-84wf8"] Dec 05 13:42:28 crc kubenswrapper[4809]: I1205 13:42:28.891881 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" path="/var/lib/kubelet/pods/b299a302-8020-4c2d-b9da-72d23d9dc042/volumes" Dec 05 13:42:29 crc kubenswrapper[4809]: I1205 13:42:29.872908 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:42:29 crc kubenswrapper[4809]: E1205 13:42:29.873492 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:42:43 crc kubenswrapper[4809]: I1205 13:42:43.872915 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:42:43 crc kubenswrapper[4809]: E1205 13:42:43.873751 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:42:56 crc kubenswrapper[4809]: I1205 13:42:56.872286 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:42:56 crc kubenswrapper[4809]: E1205 13:42:56.874532 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:43:11 crc kubenswrapper[4809]: I1205 13:43:11.872215 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:43:11 crc kubenswrapper[4809]: E1205 13:43:11.873994 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:43:26 crc kubenswrapper[4809]: I1205 13:43:26.873440 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:43:26 crc kubenswrapper[4809]: E1205 13:43:26.874477 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:43:37 crc kubenswrapper[4809]: I1205 13:43:37.873866 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:43:37 crc kubenswrapper[4809]: E1205 13:43:37.874875 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:43:47 crc kubenswrapper[4809]: I1205 13:43:47.754354 4809 scope.go:117] "RemoveContainer" containerID="0be806feae39081051529d3d435353e086d03a85766f19b29677649e74a52e2e" Dec 05 13:43:47 crc kubenswrapper[4809]: I1205 13:43:47.781891 4809 scope.go:117] "RemoveContainer" containerID="6c5f765f59f66dddb4c2eed67bce12a559b4d6afa312bb3effbc96c0111b992d" Dec 05 13:43:47 crc kubenswrapper[4809]: I1205 13:43:47.843584 4809 scope.go:117] "RemoveContainer" containerID="d53500f0846bc15721614b742796cfc8f5f3052be546dd1b3206484479a63386" Dec 05 13:43:51 crc kubenswrapper[4809]: I1205 13:43:51.872142 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:43:51 crc kubenswrapper[4809]: E1205 13:43:51.872930 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:44:02 crc kubenswrapper[4809]: I1205 13:44:02.873001 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:44:02 crc kubenswrapper[4809]: E1205 13:44:02.873929 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:44:17 crc kubenswrapper[4809]: I1205 13:44:17.872581 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:44:17 crc kubenswrapper[4809]: E1205 13:44:17.873405 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:44:28 crc kubenswrapper[4809]: I1205 13:44:28.872570 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:44:28 crc kubenswrapper[4809]: E1205 13:44:28.873650 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.043658 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hbfwc"] Dec 05 13:44:39 crc kubenswrapper[4809]: E1205 13:44:39.044748 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="extract-content" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.044767 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="extract-content" Dec 05 13:44:39 crc kubenswrapper[4809]: E1205 13:44:39.044801 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="extract-utilities" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.044810 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="extract-utilities" Dec 05 13:44:39 crc kubenswrapper[4809]: E1205 13:44:39.044828 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="registry-server" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.044836 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="registry-server" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.045052 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b299a302-8020-4c2d-b9da-72d23d9dc042" containerName="registry-server" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.046734 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.061992 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbfwc"] Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.162173 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fqtr\" (UniqueName: \"kubernetes.io/projected/8c8a83fd-0ee9-4226-a973-a1a887e70715-kube-api-access-6fqtr\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.162282 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-utilities\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.162363 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-catalog-content\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.264441 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fqtr\" (UniqueName: \"kubernetes.io/projected/8c8a83fd-0ee9-4226-a973-a1a887e70715-kube-api-access-6fqtr\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.264554 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-utilities\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.264627 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-catalog-content\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.265345 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-utilities\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.265353 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-catalog-content\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.386565 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fqtr\" (UniqueName: \"kubernetes.io/projected/8c8a83fd-0ee9-4226-a973-a1a887e70715-kube-api-access-6fqtr\") pod \"redhat-marketplace-hbfwc\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:39 crc kubenswrapper[4809]: I1205 13:44:39.669372 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:40 crc kubenswrapper[4809]: I1205 13:44:40.136970 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbfwc"] Dec 05 13:44:40 crc kubenswrapper[4809]: I1205 13:44:40.845258 4809 generic.go:334] "Generic (PLEG): container finished" podID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerID="e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba" exitCode=0 Dec 05 13:44:40 crc kubenswrapper[4809]: I1205 13:44:40.845319 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbfwc" event={"ID":"8c8a83fd-0ee9-4226-a973-a1a887e70715","Type":"ContainerDied","Data":"e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba"} Dec 05 13:44:40 crc kubenswrapper[4809]: I1205 13:44:40.845527 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbfwc" event={"ID":"8c8a83fd-0ee9-4226-a973-a1a887e70715","Type":"ContainerStarted","Data":"80063076ecb2631f67f5decfe6c1c839481ef6551c173691fdf35ae5374c8c8c"} Dec 05 13:44:40 crc kubenswrapper[4809]: I1205 13:44:40.847553 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:44:42 crc kubenswrapper[4809]: I1205 13:44:42.868338 4809 generic.go:334] "Generic (PLEG): container finished" podID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerID="f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92" exitCode=0 Dec 05 13:44:42 crc kubenswrapper[4809]: I1205 13:44:42.868433 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbfwc" event={"ID":"8c8a83fd-0ee9-4226-a973-a1a887e70715","Type":"ContainerDied","Data":"f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92"} Dec 05 13:44:43 crc kubenswrapper[4809]: I1205 13:44:43.873080 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:44:43 crc kubenswrapper[4809]: E1205 13:44:43.873657 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:44:45 crc kubenswrapper[4809]: I1205 13:44:45.908169 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbfwc" event={"ID":"8c8a83fd-0ee9-4226-a973-a1a887e70715","Type":"ContainerStarted","Data":"b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610"} Dec 05 13:44:45 crc kubenswrapper[4809]: I1205 13:44:45.938090 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hbfwc" podStartSLOduration=3.008117463 podStartE2EDuration="6.938060858s" podCreationTimestamp="2025-12-05 13:44:39 +0000 UTC" firstStartedPulling="2025-12-05 13:44:40.847318907 +0000 UTC m=+9376.238295465" lastFinishedPulling="2025-12-05 13:44:44.777262302 +0000 UTC m=+9380.168238860" observedRunningTime="2025-12-05 13:44:45.930112625 +0000 UTC m=+9381.321089203" watchObservedRunningTime="2025-12-05 13:44:45.938060858 +0000 UTC m=+9381.329037446" Dec 05 13:44:49 crc kubenswrapper[4809]: I1205 13:44:49.671065 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:49 crc kubenswrapper[4809]: I1205 13:44:49.671620 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:49 crc kubenswrapper[4809]: I1205 13:44:49.746776 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:56 crc kubenswrapper[4809]: I1205 13:44:56.871855 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:44:56 crc kubenswrapper[4809]: E1205 13:44:56.872704 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:44:59 crc kubenswrapper[4809]: I1205 13:44:59.745221 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:44:59 crc kubenswrapper[4809]: I1205 13:44:59.806006 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbfwc"] Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.048863 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hbfwc" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="registry-server" containerID="cri-o://b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610" gracePeriod=2 Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.157661 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq"] Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.160332 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.167198 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.169378 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.184186 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq"] Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.334120 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e53978e8-b25c-49f0-8bcc-a0650f45c422-config-volume\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.334696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e53978e8-b25c-49f0-8bcc-a0650f45c422-secret-volume\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.334750 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrrlx\" (UniqueName: \"kubernetes.io/projected/e53978e8-b25c-49f0-8bcc-a0650f45c422-kube-api-access-mrrlx\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.437680 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrrlx\" (UniqueName: \"kubernetes.io/projected/e53978e8-b25c-49f0-8bcc-a0650f45c422-kube-api-access-mrrlx\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.437748 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e53978e8-b25c-49f0-8bcc-a0650f45c422-config-volume\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.437929 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e53978e8-b25c-49f0-8bcc-a0650f45c422-secret-volume\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.438967 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e53978e8-b25c-49f0-8bcc-a0650f45c422-config-volume\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.448336 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e53978e8-b25c-49f0-8bcc-a0650f45c422-secret-volume\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.455157 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrrlx\" (UniqueName: \"kubernetes.io/projected/e53978e8-b25c-49f0-8bcc-a0650f45c422-kube-api-access-mrrlx\") pod \"collect-profiles-29415705-7xcdq\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.545702 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.672682 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.846973 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-utilities\") pod \"8c8a83fd-0ee9-4226-a973-a1a887e70715\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.847142 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-catalog-content\") pod \"8c8a83fd-0ee9-4226-a973-a1a887e70715\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.847206 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fqtr\" (UniqueName: \"kubernetes.io/projected/8c8a83fd-0ee9-4226-a973-a1a887e70715-kube-api-access-6fqtr\") pod \"8c8a83fd-0ee9-4226-a973-a1a887e70715\" (UID: \"8c8a83fd-0ee9-4226-a973-a1a887e70715\") " Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.848722 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-utilities" (OuterVolumeSpecName: "utilities") pod "8c8a83fd-0ee9-4226-a973-a1a887e70715" (UID: "8c8a83fd-0ee9-4226-a973-a1a887e70715"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.854782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8a83fd-0ee9-4226-a973-a1a887e70715-kube-api-access-6fqtr" (OuterVolumeSpecName: "kube-api-access-6fqtr") pod "8c8a83fd-0ee9-4226-a973-a1a887e70715" (UID: "8c8a83fd-0ee9-4226-a973-a1a887e70715"). InnerVolumeSpecName "kube-api-access-6fqtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.867350 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c8a83fd-0ee9-4226-a973-a1a887e70715" (UID: "8c8a83fd-0ee9-4226-a973-a1a887e70715"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.950295 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.950869 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a83fd-0ee9-4226-a973-a1a887e70715-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:45:00 crc kubenswrapper[4809]: I1205 13:45:00.950887 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fqtr\" (UniqueName: \"kubernetes.io/projected/8c8a83fd-0ee9-4226-a973-a1a887e70715-kube-api-access-6fqtr\") on node \"crc\" DevicePath \"\"" Dec 05 13:45:01 crc kubenswrapper[4809]: W1205 13:45:01.051872 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode53978e8_b25c_49f0_8bcc_a0650f45c422.slice/crio-8c748bdba44de446849be104e1dc61ab43e307489d15c2a69e007085bcd429f5 WatchSource:0}: Error finding container 8c748bdba44de446849be104e1dc61ab43e307489d15c2a69e007085bcd429f5: Status 404 returned error can't find the container with id 8c748bdba44de446849be104e1dc61ab43e307489d15c2a69e007085bcd429f5 Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.061509 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq"] Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.061819 4809 generic.go:334] "Generic (PLEG): container finished" podID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerID="b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610" exitCode=0 Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.061866 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbfwc" event={"ID":"8c8a83fd-0ee9-4226-a973-a1a887e70715","Type":"ContainerDied","Data":"b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610"} Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.061895 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbfwc" event={"ID":"8c8a83fd-0ee9-4226-a973-a1a887e70715","Type":"ContainerDied","Data":"80063076ecb2631f67f5decfe6c1c839481ef6551c173691fdf35ae5374c8c8c"} Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.061914 4809 scope.go:117] "RemoveContainer" containerID="b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.063481 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbfwc" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.097450 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbfwc"] Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.115951 4809 scope.go:117] "RemoveContainer" containerID="f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.119865 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbfwc"] Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.147979 4809 scope.go:117] "RemoveContainer" containerID="e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.183700 4809 scope.go:117] "RemoveContainer" containerID="b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610" Dec 05 13:45:01 crc kubenswrapper[4809]: E1205 13:45:01.184306 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610\": container with ID starting with b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610 not found: ID does not exist" containerID="b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.184357 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610"} err="failed to get container status \"b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610\": rpc error: code = NotFound desc = could not find container \"b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610\": container with ID starting with b29a00679ef020f0a4e184254e22144b7c57303292fc1d738c9ba3a9979c2610 not found: ID does not exist" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.184388 4809 scope.go:117] "RemoveContainer" containerID="f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92" Dec 05 13:45:01 crc kubenswrapper[4809]: E1205 13:45:01.184943 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92\": container with ID starting with f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92 not found: ID does not exist" containerID="f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.184998 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92"} err="failed to get container status \"f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92\": rpc error: code = NotFound desc = could not find container \"f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92\": container with ID starting with f4afd3d883c1d396a45c9bea00d1ed283715ce682cb060c6020bd3f6e72b4d92 not found: ID does not exist" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.185030 4809 scope.go:117] "RemoveContainer" containerID="e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba" Dec 05 13:45:01 crc kubenswrapper[4809]: E1205 13:45:01.185466 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba\": container with ID starting with e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba not found: ID does not exist" containerID="e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba" Dec 05 13:45:01 crc kubenswrapper[4809]: I1205 13:45:01.185489 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba"} err="failed to get container status \"e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba\": rpc error: code = NotFound desc = could not find container \"e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba\": container with ID starting with e72a170cfab8b79cda6f56df0069a5169b01d1b8e599c82a0b88c97813e513ba not found: ID does not exist" Dec 05 13:45:02 crc kubenswrapper[4809]: I1205 13:45:02.071922 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" event={"ID":"e53978e8-b25c-49f0-8bcc-a0650f45c422","Type":"ContainerStarted","Data":"8c748bdba44de446849be104e1dc61ab43e307489d15c2a69e007085bcd429f5"} Dec 05 13:45:02 crc kubenswrapper[4809]: I1205 13:45:02.900087 4809 trace.go:236] Trace[1737424143]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-cell1-server-0" (05-Dec-2025 13:45:01.728) (total time: 1171ms): Dec 05 13:45:02 crc kubenswrapper[4809]: Trace[1737424143]: [1.171963297s] [1.171963297s] END Dec 05 13:45:02 crc kubenswrapper[4809]: I1205 13:45:02.926590 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" path="/var/lib/kubelet/pods/8c8a83fd-0ee9-4226-a973-a1a887e70715/volumes" Dec 05 13:45:03 crc kubenswrapper[4809]: I1205 13:45:03.084371 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" event={"ID":"e53978e8-b25c-49f0-8bcc-a0650f45c422","Type":"ContainerStarted","Data":"1590118aa6a56aff8c69231702a562cb2019aca1f557f1123236dc3ceeab782f"} Dec 05 13:45:05 crc kubenswrapper[4809]: I1205 13:45:05.115973 4809 generic.go:334] "Generic (PLEG): container finished" podID="e53978e8-b25c-49f0-8bcc-a0650f45c422" containerID="1590118aa6a56aff8c69231702a562cb2019aca1f557f1123236dc3ceeab782f" exitCode=0 Dec 05 13:45:05 crc kubenswrapper[4809]: I1205 13:45:05.116042 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" event={"ID":"e53978e8-b25c-49f0-8bcc-a0650f45c422","Type":"ContainerDied","Data":"1590118aa6a56aff8c69231702a562cb2019aca1f557f1123236dc3ceeab782f"} Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.610597 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.780884 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrrlx\" (UniqueName: \"kubernetes.io/projected/e53978e8-b25c-49f0-8bcc-a0650f45c422-kube-api-access-mrrlx\") pod \"e53978e8-b25c-49f0-8bcc-a0650f45c422\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.781040 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e53978e8-b25c-49f0-8bcc-a0650f45c422-config-volume\") pod \"e53978e8-b25c-49f0-8bcc-a0650f45c422\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.781114 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e53978e8-b25c-49f0-8bcc-a0650f45c422-secret-volume\") pod \"e53978e8-b25c-49f0-8bcc-a0650f45c422\" (UID: \"e53978e8-b25c-49f0-8bcc-a0650f45c422\") " Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.781781 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e53978e8-b25c-49f0-8bcc-a0650f45c422-config-volume" (OuterVolumeSpecName: "config-volume") pod "e53978e8-b25c-49f0-8bcc-a0650f45c422" (UID: "e53978e8-b25c-49f0-8bcc-a0650f45c422"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.786824 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e53978e8-b25c-49f0-8bcc-a0650f45c422-kube-api-access-mrrlx" (OuterVolumeSpecName: "kube-api-access-mrrlx") pod "e53978e8-b25c-49f0-8bcc-a0650f45c422" (UID: "e53978e8-b25c-49f0-8bcc-a0650f45c422"). InnerVolumeSpecName "kube-api-access-mrrlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.787289 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53978e8-b25c-49f0-8bcc-a0650f45c422-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e53978e8-b25c-49f0-8bcc-a0650f45c422" (UID: "e53978e8-b25c-49f0-8bcc-a0650f45c422"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.883576 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e53978e8-b25c-49f0-8bcc-a0650f45c422-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.883613 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrrlx\" (UniqueName: \"kubernetes.io/projected/e53978e8-b25c-49f0-8bcc-a0650f45c422-kube-api-access-mrrlx\") on node \"crc\" DevicePath \"\"" Dec 05 13:45:06 crc kubenswrapper[4809]: I1205 13:45:06.883638 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e53978e8-b25c-49f0-8bcc-a0650f45c422-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 13:45:07 crc kubenswrapper[4809]: I1205 13:45:07.138271 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" event={"ID":"e53978e8-b25c-49f0-8bcc-a0650f45c422","Type":"ContainerDied","Data":"8c748bdba44de446849be104e1dc61ab43e307489d15c2a69e007085bcd429f5"} Dec 05 13:45:07 crc kubenswrapper[4809]: I1205 13:45:07.138714 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c748bdba44de446849be104e1dc61ab43e307489d15c2a69e007085bcd429f5" Dec 05 13:45:07 crc kubenswrapper[4809]: I1205 13:45:07.138356 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415705-7xcdq" Dec 05 13:45:07 crc kubenswrapper[4809]: I1205 13:45:07.701332 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7"] Dec 05 13:45:07 crc kubenswrapper[4809]: I1205 13:45:07.714418 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415660-qvzv7"] Dec 05 13:45:07 crc kubenswrapper[4809]: I1205 13:45:07.872784 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:45:07 crc kubenswrapper[4809]: E1205 13:45:07.873165 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:45:08 crc kubenswrapper[4809]: I1205 13:45:08.888507 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="886f4dbc-fe92-423b-aede-f14c3e321095" path="/var/lib/kubelet/pods/886f4dbc-fe92-423b-aede-f14c3e321095/volumes" Dec 05 13:45:19 crc kubenswrapper[4809]: I1205 13:45:19.872094 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:45:19 crc kubenswrapper[4809]: E1205 13:45:19.874735 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:45:30 crc kubenswrapper[4809]: I1205 13:45:30.872815 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:45:30 crc kubenswrapper[4809]: E1205 13:45:30.873958 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:45:42 crc kubenswrapper[4809]: I1205 13:45:42.872177 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:45:42 crc kubenswrapper[4809]: E1205 13:45:42.872958 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:45:47 crc kubenswrapper[4809]: I1205 13:45:47.918591 4809 scope.go:117] "RemoveContainer" containerID="503b339a61e7c04302fb430ca56f2aa32a7863d49235444049daf6ab7a4d2808" Dec 05 13:45:55 crc kubenswrapper[4809]: I1205 13:45:55.872289 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:45:55 crc kubenswrapper[4809]: E1205 13:45:55.873376 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:46:08 crc kubenswrapper[4809]: I1205 13:46:08.873164 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:46:08 crc kubenswrapper[4809]: E1205 13:46:08.874140 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:46:20 crc kubenswrapper[4809]: I1205 13:46:20.873102 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:46:20 crc kubenswrapper[4809]: E1205 13:46:20.875815 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:46:35 crc kubenswrapper[4809]: I1205 13:46:35.872685 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:46:35 crc kubenswrapper[4809]: E1205 13:46:35.873511 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:46:46 crc kubenswrapper[4809]: I1205 13:46:46.872868 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:46:47 crc kubenswrapper[4809]: I1205 13:46:47.295471 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"ed4f444a78eb4a4541d40910e0e3d50538ccec421a579187e7645cf38d2ff4f0"} Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.631331 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zbhnl"] Dec 05 13:47:56 crc kubenswrapper[4809]: E1205 13:47:56.632324 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="extract-utilities" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.632342 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="extract-utilities" Dec 05 13:47:56 crc kubenswrapper[4809]: E1205 13:47:56.632386 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="extract-content" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.632393 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="extract-content" Dec 05 13:47:56 crc kubenswrapper[4809]: E1205 13:47:56.632411 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53978e8-b25c-49f0-8bcc-a0650f45c422" containerName="collect-profiles" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.632418 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53978e8-b25c-49f0-8bcc-a0650f45c422" containerName="collect-profiles" Dec 05 13:47:56 crc kubenswrapper[4809]: E1205 13:47:56.632426 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="registry-server" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.632432 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="registry-server" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.632674 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c8a83fd-0ee9-4226-a973-a1a887e70715" containerName="registry-server" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.632689 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53978e8-b25c-49f0-8bcc-a0650f45c422" containerName="collect-profiles" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.634614 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.651460 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zbhnl"] Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.741181 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-utilities\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.741256 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r75gc\" (UniqueName: \"kubernetes.io/projected/0bc5911b-8b80-438c-b1c9-fe88a936d943-kube-api-access-r75gc\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.741427 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-catalog-content\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.844289 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-catalog-content\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.844367 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-utilities\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.844403 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r75gc\" (UniqueName: \"kubernetes.io/projected/0bc5911b-8b80-438c-b1c9-fe88a936d943-kube-api-access-r75gc\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.845112 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-catalog-content\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.845390 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-utilities\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.864824 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r75gc\" (UniqueName: \"kubernetes.io/projected/0bc5911b-8b80-438c-b1c9-fe88a936d943-kube-api-access-r75gc\") pod \"redhat-operators-zbhnl\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:56 crc kubenswrapper[4809]: I1205 13:47:56.964714 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:47:57 crc kubenswrapper[4809]: I1205 13:47:57.445147 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zbhnl"] Dec 05 13:47:58 crc kubenswrapper[4809]: I1205 13:47:58.090932 4809 generic.go:334] "Generic (PLEG): container finished" podID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerID="c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae" exitCode=0 Dec 05 13:47:58 crc kubenswrapper[4809]: I1205 13:47:58.091002 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbhnl" event={"ID":"0bc5911b-8b80-438c-b1c9-fe88a936d943","Type":"ContainerDied","Data":"c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae"} Dec 05 13:47:58 crc kubenswrapper[4809]: I1205 13:47:58.091243 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbhnl" event={"ID":"0bc5911b-8b80-438c-b1c9-fe88a936d943","Type":"ContainerStarted","Data":"c2023fcf1c81ec7c7c37e8dfd633e2ecdd20bf7153dce0db6ad3a7203191b00f"} Dec 05 13:48:00 crc kubenswrapper[4809]: I1205 13:48:00.117731 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbhnl" event={"ID":"0bc5911b-8b80-438c-b1c9-fe88a936d943","Type":"ContainerStarted","Data":"c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26"} Dec 05 13:48:02 crc kubenswrapper[4809]: I1205 13:48:02.144478 4809 generic.go:334] "Generic (PLEG): container finished" podID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerID="c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26" exitCode=0 Dec 05 13:48:02 crc kubenswrapper[4809]: I1205 13:48:02.144655 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbhnl" event={"ID":"0bc5911b-8b80-438c-b1c9-fe88a936d943","Type":"ContainerDied","Data":"c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26"} Dec 05 13:48:03 crc kubenswrapper[4809]: I1205 13:48:03.160896 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbhnl" event={"ID":"0bc5911b-8b80-438c-b1c9-fe88a936d943","Type":"ContainerStarted","Data":"82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda"} Dec 05 13:48:03 crc kubenswrapper[4809]: I1205 13:48:03.188859 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zbhnl" podStartSLOduration=2.738941954 podStartE2EDuration="7.188840085s" podCreationTimestamp="2025-12-05 13:47:56 +0000 UTC" firstStartedPulling="2025-12-05 13:47:58.092800072 +0000 UTC m=+9573.483776630" lastFinishedPulling="2025-12-05 13:48:02.542698203 +0000 UTC m=+9577.933674761" observedRunningTime="2025-12-05 13:48:03.182883205 +0000 UTC m=+9578.573859803" watchObservedRunningTime="2025-12-05 13:48:03.188840085 +0000 UTC m=+9578.579816643" Dec 05 13:48:04 crc kubenswrapper[4809]: I1205 13:48:04.172526 4809 generic.go:334] "Generic (PLEG): container finished" podID="2959aa37-7f80-4b96-9284-c508595cce86" containerID="0558eee863b24d0be7642f71c019a3f9da7c11ce41aeb81e15f189c59e35dd0b" exitCode=0 Dec 05 13:48:04 crc kubenswrapper[4809]: I1205 13:48:04.172714 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" event={"ID":"2959aa37-7f80-4b96-9284-c508595cce86","Type":"ContainerDied","Data":"0558eee863b24d0be7642f71c019a3f9da7c11ce41aeb81e15f189c59e35dd0b"} Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.721012 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.870888 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-0\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871042 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nqp7\" (UniqueName: \"kubernetes.io/projected/2959aa37-7f80-4b96-9284-c508595cce86-kube-api-access-4nqp7\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871128 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-1\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871192 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-inventory\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871225 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-1\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871256 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-combined-ca-bundle\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871339 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-0\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871381 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-1\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871422 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ceph\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871472 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-0\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.871527 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ssh-key\") pod \"2959aa37-7f80-4b96-9284-c508595cce86\" (UID: \"2959aa37-7f80-4b96-9284-c508595cce86\") " Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.877861 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ceph" (OuterVolumeSpecName: "ceph") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.899807 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2959aa37-7f80-4b96-9284-c508595cce86-kube-api-access-4nqp7" (OuterVolumeSpecName: "kube-api-access-4nqp7") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "kube-api-access-4nqp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.901270 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.905494 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.911887 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-inventory" (OuterVolumeSpecName: "inventory") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.911956 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.912480 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.913431 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.917058 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.917547 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.932173 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2959aa37-7f80-4b96-9284-c508595cce86" (UID: "2959aa37-7f80-4b96-9284-c508595cce86"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975258 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nqp7\" (UniqueName: \"kubernetes.io/projected/2959aa37-7f80-4b96-9284-c508595cce86-kube-api-access-4nqp7\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975293 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975305 4809 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975315 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975325 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975335 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2959aa37-7f80-4b96-9284-c508595cce86-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975346 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975355 4809 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975364 4809 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975373 4809 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:05 crc kubenswrapper[4809]: I1205 13:48:05.975383 4809 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2959aa37-7f80-4b96-9284-c508595cce86-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:06 crc kubenswrapper[4809]: I1205 13:48:06.199466 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" event={"ID":"2959aa37-7f80-4b96-9284-c508595cce86","Type":"ContainerDied","Data":"f19804925f034716756433db8dd34b19f1ca345389f9aded2aa49bafa64495ad"} Dec 05 13:48:06 crc kubenswrapper[4809]: I1205 13:48:06.199536 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f19804925f034716756433db8dd34b19f1ca345389f9aded2aa49bafa64495ad" Dec 05 13:48:06 crc kubenswrapper[4809]: I1205 13:48:06.199616 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv" Dec 05 13:48:06 crc kubenswrapper[4809]: I1205 13:48:06.965413 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:48:06 crc kubenswrapper[4809]: I1205 13:48:06.965533 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:48:08 crc kubenswrapper[4809]: I1205 13:48:08.045763 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zbhnl" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="registry-server" probeResult="failure" output=< Dec 05 13:48:08 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 13:48:08 crc kubenswrapper[4809]: > Dec 05 13:48:17 crc kubenswrapper[4809]: I1205 13:48:17.014743 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:48:17 crc kubenswrapper[4809]: I1205 13:48:17.070672 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:48:17 crc kubenswrapper[4809]: I1205 13:48:17.580170 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zbhnl"] Dec 05 13:48:18 crc kubenswrapper[4809]: I1205 13:48:18.348494 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zbhnl" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="registry-server" containerID="cri-o://82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda" gracePeriod=2 Dec 05 13:48:18 crc kubenswrapper[4809]: I1205 13:48:18.914421 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:48:18 crc kubenswrapper[4809]: I1205 13:48:18.991091 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r75gc\" (UniqueName: \"kubernetes.io/projected/0bc5911b-8b80-438c-b1c9-fe88a936d943-kube-api-access-r75gc\") pod \"0bc5911b-8b80-438c-b1c9-fe88a936d943\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " Dec 05 13:48:18 crc kubenswrapper[4809]: I1205 13:48:18.991652 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-utilities\") pod \"0bc5911b-8b80-438c-b1c9-fe88a936d943\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " Dec 05 13:48:18 crc kubenswrapper[4809]: I1205 13:48:18.991721 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-catalog-content\") pod \"0bc5911b-8b80-438c-b1c9-fe88a936d943\" (UID: \"0bc5911b-8b80-438c-b1c9-fe88a936d943\") " Dec 05 13:48:18 crc kubenswrapper[4809]: I1205 13:48:18.992447 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-utilities" (OuterVolumeSpecName: "utilities") pod "0bc5911b-8b80-438c-b1c9-fe88a936d943" (UID: "0bc5911b-8b80-438c-b1c9-fe88a936d943"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:48:18 crc kubenswrapper[4809]: I1205 13:48:18.996933 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc5911b-8b80-438c-b1c9-fe88a936d943-kube-api-access-r75gc" (OuterVolumeSpecName: "kube-api-access-r75gc") pod "0bc5911b-8b80-438c-b1c9-fe88a936d943" (UID: "0bc5911b-8b80-438c-b1c9-fe88a936d943"). InnerVolumeSpecName "kube-api-access-r75gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.094125 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.094167 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r75gc\" (UniqueName: \"kubernetes.io/projected/0bc5911b-8b80-438c-b1c9-fe88a936d943-kube-api-access-r75gc\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.109333 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bc5911b-8b80-438c-b1c9-fe88a936d943" (UID: "0bc5911b-8b80-438c-b1c9-fe88a936d943"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.197438 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc5911b-8b80-438c-b1c9-fe88a936d943-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.360815 4809 generic.go:334] "Generic (PLEG): container finished" podID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerID="82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda" exitCode=0 Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.360865 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbhnl" event={"ID":"0bc5911b-8b80-438c-b1c9-fe88a936d943","Type":"ContainerDied","Data":"82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda"} Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.360891 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbhnl" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.360919 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbhnl" event={"ID":"0bc5911b-8b80-438c-b1c9-fe88a936d943","Type":"ContainerDied","Data":"c2023fcf1c81ec7c7c37e8dfd633e2ecdd20bf7153dce0db6ad3a7203191b00f"} Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.360942 4809 scope.go:117] "RemoveContainer" containerID="82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.385336 4809 scope.go:117] "RemoveContainer" containerID="c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.410724 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zbhnl"] Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.420875 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zbhnl"] Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.422031 4809 scope.go:117] "RemoveContainer" containerID="c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.474607 4809 scope.go:117] "RemoveContainer" containerID="82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda" Dec 05 13:48:19 crc kubenswrapper[4809]: E1205 13:48:19.475216 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda\": container with ID starting with 82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda not found: ID does not exist" containerID="82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.475259 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda"} err="failed to get container status \"82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda\": rpc error: code = NotFound desc = could not find container \"82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda\": container with ID starting with 82d1880176dc22222fa047cfd5d33ed665612f0e28fd4f6e2380cf59ada85bda not found: ID does not exist" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.475287 4809 scope.go:117] "RemoveContainer" containerID="c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26" Dec 05 13:48:19 crc kubenswrapper[4809]: E1205 13:48:19.475748 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26\": container with ID starting with c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26 not found: ID does not exist" containerID="c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.475790 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26"} err="failed to get container status \"c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26\": rpc error: code = NotFound desc = could not find container \"c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26\": container with ID starting with c155334bb7f147ad1c5fda7e5764964e3e92a65751a3906aaa5c0b56a889cd26 not found: ID does not exist" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.475815 4809 scope.go:117] "RemoveContainer" containerID="c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae" Dec 05 13:48:19 crc kubenswrapper[4809]: E1205 13:48:19.476159 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae\": container with ID starting with c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae not found: ID does not exist" containerID="c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae" Dec 05 13:48:19 crc kubenswrapper[4809]: I1205 13:48:19.476194 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae"} err="failed to get container status \"c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae\": rpc error: code = NotFound desc = could not find container \"c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae\": container with ID starting with c8403db41cdc929bc3ea52bcee10e202b933ab72f8a84a1da631d41a4b0d5fae not found: ID does not exist" Dec 05 13:48:20 crc kubenswrapper[4809]: I1205 13:48:20.890989 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" path="/var/lib/kubelet/pods/0bc5911b-8b80-438c-b1c9-fe88a936d943/volumes" Dec 05 13:49:14 crc kubenswrapper[4809]: I1205 13:49:14.046886 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:49:14 crc kubenswrapper[4809]: I1205 13:49:14.047464 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.290044 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h99hn"] Dec 05 13:49:30 crc kubenswrapper[4809]: E1205 13:49:30.291172 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="registry-server" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.291190 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="registry-server" Dec 05 13:49:30 crc kubenswrapper[4809]: E1205 13:49:30.291223 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="extract-content" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.291231 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="extract-content" Dec 05 13:49:30 crc kubenswrapper[4809]: E1205 13:49:30.291242 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="extract-utilities" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.291289 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="extract-utilities" Dec 05 13:49:30 crc kubenswrapper[4809]: E1205 13:49:30.291333 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2959aa37-7f80-4b96-9284-c508595cce86" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.291342 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="2959aa37-7f80-4b96-9284-c508595cce86" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.291585 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="2959aa37-7f80-4b96-9284-c508595cce86" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.291612 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc5911b-8b80-438c-b1c9-fe88a936d943" containerName="registry-server" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.302527 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.329849 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-catalog-content\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.331994 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd72v\" (UniqueName: \"kubernetes.io/projected/a54df50c-eabf-4052-a8a2-9f6382894cfe-kube-api-access-rd72v\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.333471 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-utilities\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.359437 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h99hn"] Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.435669 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-catalog-content\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.435813 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd72v\" (UniqueName: \"kubernetes.io/projected/a54df50c-eabf-4052-a8a2-9f6382894cfe-kube-api-access-rd72v\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.435940 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-utilities\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.436199 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-catalog-content\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.436480 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-utilities\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.455504 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd72v\" (UniqueName: \"kubernetes.io/projected/a54df50c-eabf-4052-a8a2-9f6382894cfe-kube-api-access-rd72v\") pod \"community-operators-h99hn\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:30 crc kubenswrapper[4809]: I1205 13:49:30.632537 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:31 crc kubenswrapper[4809]: I1205 13:49:31.191158 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h99hn"] Dec 05 13:49:32 crc kubenswrapper[4809]: I1205 13:49:32.182726 4809 generic.go:334] "Generic (PLEG): container finished" podID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerID="ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96" exitCode=0 Dec 05 13:49:32 crc kubenswrapper[4809]: I1205 13:49:32.182833 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h99hn" event={"ID":"a54df50c-eabf-4052-a8a2-9f6382894cfe","Type":"ContainerDied","Data":"ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96"} Dec 05 13:49:32 crc kubenswrapper[4809]: I1205 13:49:32.183102 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h99hn" event={"ID":"a54df50c-eabf-4052-a8a2-9f6382894cfe","Type":"ContainerStarted","Data":"fe82bade660e45f093b2fa024f964a59131149a22fa38f193f4d35eb61463f1a"} Dec 05 13:49:33 crc kubenswrapper[4809]: I1205 13:49:33.203900 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h99hn" event={"ID":"a54df50c-eabf-4052-a8a2-9f6382894cfe","Type":"ContainerStarted","Data":"d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed"} Dec 05 13:49:34 crc kubenswrapper[4809]: I1205 13:49:34.216284 4809 generic.go:334] "Generic (PLEG): container finished" podID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerID="d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed" exitCode=0 Dec 05 13:49:34 crc kubenswrapper[4809]: I1205 13:49:34.216606 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h99hn" event={"ID":"a54df50c-eabf-4052-a8a2-9f6382894cfe","Type":"ContainerDied","Data":"d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed"} Dec 05 13:49:35 crc kubenswrapper[4809]: I1205 13:49:35.231204 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h99hn" event={"ID":"a54df50c-eabf-4052-a8a2-9f6382894cfe","Type":"ContainerStarted","Data":"fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564"} Dec 05 13:49:35 crc kubenswrapper[4809]: I1205 13:49:35.252103 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h99hn" podStartSLOduration=2.7838573970000002 podStartE2EDuration="5.252082214s" podCreationTimestamp="2025-12-05 13:49:30 +0000 UTC" firstStartedPulling="2025-12-05 13:49:32.184583852 +0000 UTC m=+9667.575560410" lastFinishedPulling="2025-12-05 13:49:34.652808669 +0000 UTC m=+9670.043785227" observedRunningTime="2025-12-05 13:49:35.250324707 +0000 UTC m=+9670.641301265" watchObservedRunningTime="2025-12-05 13:49:35.252082214 +0000 UTC m=+9670.643058772" Dec 05 13:49:40 crc kubenswrapper[4809]: I1205 13:49:40.633431 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:40 crc kubenswrapper[4809]: I1205 13:49:40.633913 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:40 crc kubenswrapper[4809]: I1205 13:49:40.692534 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:41 crc kubenswrapper[4809]: I1205 13:49:41.378004 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:41 crc kubenswrapper[4809]: I1205 13:49:41.447764 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h99hn"] Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.310343 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h99hn" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="registry-server" containerID="cri-o://fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564" gracePeriod=2 Dec 05 13:49:43 crc kubenswrapper[4809]: E1205 13:49:43.516933 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda54df50c_eabf_4052_a8a2_9f6382894cfe.slice/crio-fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda54df50c_eabf_4052_a8a2_9f6382894cfe.slice/crio-conmon-fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564.scope\": RecentStats: unable to find data in memory cache]" Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.810651 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.942603 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-catalog-content\") pod \"a54df50c-eabf-4052-a8a2-9f6382894cfe\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.942701 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd72v\" (UniqueName: \"kubernetes.io/projected/a54df50c-eabf-4052-a8a2-9f6382894cfe-kube-api-access-rd72v\") pod \"a54df50c-eabf-4052-a8a2-9f6382894cfe\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.943106 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-utilities\") pod \"a54df50c-eabf-4052-a8a2-9f6382894cfe\" (UID: \"a54df50c-eabf-4052-a8a2-9f6382894cfe\") " Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.944139 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-utilities" (OuterVolumeSpecName: "utilities") pod "a54df50c-eabf-4052-a8a2-9f6382894cfe" (UID: "a54df50c-eabf-4052-a8a2-9f6382894cfe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.944301 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:49:43 crc kubenswrapper[4809]: I1205 13:49:43.957879 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54df50c-eabf-4052-a8a2-9f6382894cfe-kube-api-access-rd72v" (OuterVolumeSpecName: "kube-api-access-rd72v") pod "a54df50c-eabf-4052-a8a2-9f6382894cfe" (UID: "a54df50c-eabf-4052-a8a2-9f6382894cfe"). InnerVolumeSpecName "kube-api-access-rd72v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.023285 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a54df50c-eabf-4052-a8a2-9f6382894cfe" (UID: "a54df50c-eabf-4052-a8a2-9f6382894cfe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.046066 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.046121 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.048785 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a54df50c-eabf-4052-a8a2-9f6382894cfe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.048859 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd72v\" (UniqueName: \"kubernetes.io/projected/a54df50c-eabf-4052-a8a2-9f6382894cfe-kube-api-access-rd72v\") on node \"crc\" DevicePath \"\"" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.330883 4809 generic.go:334] "Generic (PLEG): container finished" podID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerID="fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564" exitCode=0 Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.330946 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h99hn" event={"ID":"a54df50c-eabf-4052-a8a2-9f6382894cfe","Type":"ContainerDied","Data":"fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564"} Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.330995 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h99hn" event={"ID":"a54df50c-eabf-4052-a8a2-9f6382894cfe","Type":"ContainerDied","Data":"fe82bade660e45f093b2fa024f964a59131149a22fa38f193f4d35eb61463f1a"} Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.331020 4809 scope.go:117] "RemoveContainer" containerID="fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.331031 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h99hn" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.379514 4809 scope.go:117] "RemoveContainer" containerID="d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.402785 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h99hn"] Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.407982 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h99hn"] Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.436476 4809 scope.go:117] "RemoveContainer" containerID="ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.457510 4809 scope.go:117] "RemoveContainer" containerID="fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564" Dec 05 13:49:44 crc kubenswrapper[4809]: E1205 13:49:44.458202 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564\": container with ID starting with fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564 not found: ID does not exist" containerID="fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.458240 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564"} err="failed to get container status \"fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564\": rpc error: code = NotFound desc = could not find container \"fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564\": container with ID starting with fe33ee34425a400f51d7248c3312bf552566d886173884f299a539a1e2aac564 not found: ID does not exist" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.458267 4809 scope.go:117] "RemoveContainer" containerID="d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed" Dec 05 13:49:44 crc kubenswrapper[4809]: E1205 13:49:44.458844 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed\": container with ID starting with d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed not found: ID does not exist" containerID="d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.458873 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed"} err="failed to get container status \"d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed\": rpc error: code = NotFound desc = could not find container \"d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed\": container with ID starting with d2cd69fab6118264d17fbb15dbfb5a3b886397be41cdafd5e1d0269a09a019ed not found: ID does not exist" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.458887 4809 scope.go:117] "RemoveContainer" containerID="ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96" Dec 05 13:49:44 crc kubenswrapper[4809]: E1205 13:49:44.459209 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96\": container with ID starting with ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96 not found: ID does not exist" containerID="ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.459258 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96"} err="failed to get container status \"ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96\": rpc error: code = NotFound desc = could not find container \"ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96\": container with ID starting with ad6d15610845e57fee8e7d7d5308b0b46f07b9be05a64b1b3141d161a74d7a96 not found: ID does not exist" Dec 05 13:49:44 crc kubenswrapper[4809]: I1205 13:49:44.888485 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" path="/var/lib/kubelet/pods/a54df50c-eabf-4052-a8a2-9f6382894cfe/volumes" Dec 05 13:50:04 crc kubenswrapper[4809]: I1205 13:50:04.545583 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 13:50:04 crc kubenswrapper[4809]: I1205 13:50:04.546338 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="1ea55d60-dde6-4172-9b5b-9b05bf610715" containerName="adoption" containerID="cri-o://36ac6df91e5181ef4f903b1a6162b47ab4ec3687b00ace0c8b1582ff7097f1f0" gracePeriod=30 Dec 05 13:50:14 crc kubenswrapper[4809]: I1205 13:50:14.046372 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:50:14 crc kubenswrapper[4809]: I1205 13:50:14.046982 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:50:14 crc kubenswrapper[4809]: I1205 13:50:14.047051 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:50:14 crc kubenswrapper[4809]: I1205 13:50:14.048052 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed4f444a78eb4a4541d40910e0e3d50538ccec421a579187e7645cf38d2ff4f0"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:50:14 crc kubenswrapper[4809]: I1205 13:50:14.048128 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://ed4f444a78eb4a4541d40910e0e3d50538ccec421a579187e7645cf38d2ff4f0" gracePeriod=600 Dec 05 13:50:15 crc kubenswrapper[4809]: I1205 13:50:15.716183 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="ed4f444a78eb4a4541d40910e0e3d50538ccec421a579187e7645cf38d2ff4f0" exitCode=0 Dec 05 13:50:15 crc kubenswrapper[4809]: I1205 13:50:15.716283 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"ed4f444a78eb4a4541d40910e0e3d50538ccec421a579187e7645cf38d2ff4f0"} Dec 05 13:50:15 crc kubenswrapper[4809]: I1205 13:50:15.716703 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55"} Dec 05 13:50:15 crc kubenswrapper[4809]: I1205 13:50:15.716735 4809 scope.go:117] "RemoveContainer" containerID="c022143ebe574aa8c8a2d52c43a3d5b7e8e343f1a2a8f398a9035cda7818e103" Dec 05 13:50:34 crc kubenswrapper[4809]: I1205 13:50:34.933818 4809 generic.go:334] "Generic (PLEG): container finished" podID="1ea55d60-dde6-4172-9b5b-9b05bf610715" containerID="36ac6df91e5181ef4f903b1a6162b47ab4ec3687b00ace0c8b1582ff7097f1f0" exitCode=137 Dec 05 13:50:34 crc kubenswrapper[4809]: I1205 13:50:34.933904 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1ea55d60-dde6-4172-9b5b-9b05bf610715","Type":"ContainerDied","Data":"36ac6df91e5181ef4f903b1a6162b47ab4ec3687b00ace0c8b1582ff7097f1f0"} Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.124984 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.209527 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\") pod \"1ea55d60-dde6-4172-9b5b-9b05bf610715\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.209956 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb8km\" (UniqueName: \"kubernetes.io/projected/1ea55d60-dde6-4172-9b5b-9b05bf610715-kube-api-access-bb8km\") pod \"1ea55d60-dde6-4172-9b5b-9b05bf610715\" (UID: \"1ea55d60-dde6-4172-9b5b-9b05bf610715\") " Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.226866 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea55d60-dde6-4172-9b5b-9b05bf610715-kube-api-access-bb8km" (OuterVolumeSpecName: "kube-api-access-bb8km") pod "1ea55d60-dde6-4172-9b5b-9b05bf610715" (UID: "1ea55d60-dde6-4172-9b5b-9b05bf610715"). InnerVolumeSpecName "kube-api-access-bb8km". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.312621 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb8km\" (UniqueName: \"kubernetes.io/projected/1ea55d60-dde6-4172-9b5b-9b05bf610715-kube-api-access-bb8km\") on node \"crc\" DevicePath \"\"" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.422461 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7" (OuterVolumeSpecName: "mariadb-data") pod "1ea55d60-dde6-4172-9b5b-9b05bf610715" (UID: "1ea55d60-dde6-4172-9b5b-9b05bf610715"). InnerVolumeSpecName "pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.518059 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\") on node \"crc\" " Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.559374 4809 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.559562 4809 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7") on node "crc" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.619667 4809 reconciler_common.go:293] "Volume detached for volume \"pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bbea020a-5d9b-49e3-b1f4-1d55d9a971c7\") on node \"crc\" DevicePath \"\"" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.950332 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1ea55d60-dde6-4172-9b5b-9b05bf610715","Type":"ContainerDied","Data":"5bec4577a4a5d51937a4ebc32503c804c2e90100b9d66378a3657adff0cd525a"} Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.951597 4809 scope.go:117] "RemoveContainer" containerID="36ac6df91e5181ef4f903b1a6162b47ab4ec3687b00ace0c8b1582ff7097f1f0" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.950421 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 05 13:50:35 crc kubenswrapper[4809]: I1205 13:50:35.992053 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 13:50:36 crc kubenswrapper[4809]: I1205 13:50:36.003588 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 05 13:50:36 crc kubenswrapper[4809]: I1205 13:50:36.620877 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 13:50:36 crc kubenswrapper[4809]: I1205 13:50:36.621378 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="c7fc85c2-1679-439c-ae19-9644f747c625" containerName="adoption" containerID="cri-o://09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4" gracePeriod=30 Dec 05 13:50:36 crc kubenswrapper[4809]: I1205 13:50:36.886876 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea55d60-dde6-4172-9b5b-9b05bf610715" path="/var/lib/kubelet/pods/1ea55d60-dde6-4172-9b5b-9b05bf610715/volumes" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.065499 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.128737 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvhfq\" (UniqueName: \"kubernetes.io/projected/c7fc85c2-1679-439c-ae19-9644f747c625-kube-api-access-vvhfq\") pod \"c7fc85c2-1679-439c-ae19-9644f747c625\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.129418 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\") pod \"c7fc85c2-1679-439c-ae19-9644f747c625\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.129759 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/c7fc85c2-1679-439c-ae19-9644f747c625-ovn-data-cert\") pod \"c7fc85c2-1679-439c-ae19-9644f747c625\" (UID: \"c7fc85c2-1679-439c-ae19-9644f747c625\") " Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.136919 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7fc85c2-1679-439c-ae19-9644f747c625-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "c7fc85c2-1679-439c-ae19-9644f747c625" (UID: "c7fc85c2-1679-439c-ae19-9644f747c625"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.138481 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7fc85c2-1679-439c-ae19-9644f747c625-kube-api-access-vvhfq" (OuterVolumeSpecName: "kube-api-access-vvhfq") pod "c7fc85c2-1679-439c-ae19-9644f747c625" (UID: "c7fc85c2-1679-439c-ae19-9644f747c625"). InnerVolumeSpecName "kube-api-access-vvhfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.151443 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb" (OuterVolumeSpecName: "ovn-data") pod "c7fc85c2-1679-439c-ae19-9644f747c625" (UID: "c7fc85c2-1679-439c-ae19-9644f747c625"). InnerVolumeSpecName "pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.232903 4809 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/c7fc85c2-1679-439c-ae19-9644f747c625-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.232960 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvhfq\" (UniqueName: \"kubernetes.io/projected/c7fc85c2-1679-439c-ae19-9644f747c625-kube-api-access-vvhfq\") on node \"crc\" DevicePath \"\"" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.233012 4809 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\") on node \"crc\" " Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.259108 4809 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.259265 4809 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb") on node "crc" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.284399 4809 generic.go:334] "Generic (PLEG): container finished" podID="c7fc85c2-1679-439c-ae19-9644f747c625" containerID="09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4" exitCode=137 Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.284452 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"c7fc85c2-1679-439c-ae19-9644f747c625","Type":"ContainerDied","Data":"09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4"} Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.284486 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"c7fc85c2-1679-439c-ae19-9644f747c625","Type":"ContainerDied","Data":"0da929ccf27c361136eb350442252629e10ddba1d089b918f7d3589c7a10e69e"} Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.284505 4809 scope.go:117] "RemoveContainer" containerID="09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.284511 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.328012 4809 scope.go:117] "RemoveContainer" containerID="09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4" Dec 05 13:51:07 crc kubenswrapper[4809]: E1205 13:51:07.328646 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4\": container with ID starting with 09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4 not found: ID does not exist" containerID="09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.328735 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4"} err="failed to get container status \"09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4\": rpc error: code = NotFound desc = could not find container \"09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4\": container with ID starting with 09fb201029e92825f3f4dee23612b4f43ee89030b8db82e69f0ae27d4894e3d4 not found: ID does not exist" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.333191 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.335121 4809 reconciler_common.go:293] "Volume detached for volume \"pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5cf60c29-0bb7-4ef2-96b8-4beade7f56bb\") on node \"crc\" DevicePath \"\"" Dec 05 13:51:07 crc kubenswrapper[4809]: I1205 13:51:07.346474 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 05 13:51:08 crc kubenswrapper[4809]: I1205 13:51:08.885127 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7fc85c2-1679-439c-ae19-9644f747c625" path="/var/lib/kubelet/pods/c7fc85c2-1679-439c-ae19-9644f747c625/volumes" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.574465 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wpqvz/must-gather-pc8r6"] Dec 05 13:52:13 crc kubenswrapper[4809]: E1205 13:52:13.575386 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea55d60-dde6-4172-9b5b-9b05bf610715" containerName="adoption" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575399 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea55d60-dde6-4172-9b5b-9b05bf610715" containerName="adoption" Dec 05 13:52:13 crc kubenswrapper[4809]: E1205 13:52:13.575416 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="extract-utilities" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575425 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="extract-utilities" Dec 05 13:52:13 crc kubenswrapper[4809]: E1205 13:52:13.575437 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="extract-content" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575443 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="extract-content" Dec 05 13:52:13 crc kubenswrapper[4809]: E1205 13:52:13.575458 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="registry-server" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575463 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="registry-server" Dec 05 13:52:13 crc kubenswrapper[4809]: E1205 13:52:13.575481 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7fc85c2-1679-439c-ae19-9644f747c625" containerName="adoption" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575487 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7fc85c2-1679-439c-ae19-9644f747c625" containerName="adoption" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575709 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea55d60-dde6-4172-9b5b-9b05bf610715" containerName="adoption" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575723 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7fc85c2-1679-439c-ae19-9644f747c625" containerName="adoption" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.575740 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54df50c-eabf-4052-a8a2-9f6382894cfe" containerName="registry-server" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.576879 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.579347 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wpqvz"/"kube-root-ca.crt" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.579476 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wpqvz"/"default-dockercfg-fm4b8" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.580328 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wpqvz"/"openshift-service-ca.crt" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.585165 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wpqvz/must-gather-pc8r6"] Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.606582 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-must-gather-output\") pod \"must-gather-pc8r6\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.606779 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d74qw\" (UniqueName: \"kubernetes.io/projected/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-kube-api-access-d74qw\") pod \"must-gather-pc8r6\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.708752 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d74qw\" (UniqueName: \"kubernetes.io/projected/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-kube-api-access-d74qw\") pod \"must-gather-pc8r6\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.708886 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-must-gather-output\") pod \"must-gather-pc8r6\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.709417 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-must-gather-output\") pod \"must-gather-pc8r6\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.733845 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d74qw\" (UniqueName: \"kubernetes.io/projected/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-kube-api-access-d74qw\") pod \"must-gather-pc8r6\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:13 crc kubenswrapper[4809]: I1205 13:52:13.898024 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 13:52:14 crc kubenswrapper[4809]: I1205 13:52:14.383024 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:52:14 crc kubenswrapper[4809]: I1205 13:52:14.389247 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wpqvz/must-gather-pc8r6"] Dec 05 13:52:15 crc kubenswrapper[4809]: I1205 13:52:15.044314 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" event={"ID":"b7e68161-aadf-4fcd-bfe0-f254200d4e5b","Type":"ContainerStarted","Data":"9ee34726abdb810aaf30c71f6ae733a67f7d0ede115a6b6ce89744c7c8ad7afe"} Dec 05 13:52:19 crc kubenswrapper[4809]: I1205 13:52:19.082191 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" event={"ID":"b7e68161-aadf-4fcd-bfe0-f254200d4e5b","Type":"ContainerStarted","Data":"a6a64210cf5c442ee8a44a35ff8bda9f12a6f87911ee8eb4650b647bbc0cb0d7"} Dec 05 13:52:20 crc kubenswrapper[4809]: I1205 13:52:20.096826 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" event={"ID":"b7e68161-aadf-4fcd-bfe0-f254200d4e5b","Type":"ContainerStarted","Data":"de2804e07062b4cd8939735fc30028607b439b7ef61115aedc2e63e7f9bbbebd"} Dec 05 13:52:20 crc kubenswrapper[4809]: I1205 13:52:20.121645 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" podStartSLOduration=2.882573733 podStartE2EDuration="7.121604459s" podCreationTimestamp="2025-12-05 13:52:13 +0000 UTC" firstStartedPulling="2025-12-05 13:52:14.382844485 +0000 UTC m=+9829.773821043" lastFinishedPulling="2025-12-05 13:52:18.621875211 +0000 UTC m=+9834.012851769" observedRunningTime="2025-12-05 13:52:20.111248701 +0000 UTC m=+9835.502225299" watchObservedRunningTime="2025-12-05 13:52:20.121604459 +0000 UTC m=+9835.512581017" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.067838 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wpqvz/crc-debug-tkvbd"] Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.069848 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.236216 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e28b10f3-865f-4197-a095-f559444a1d2c-host\") pod \"crc-debug-tkvbd\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.236405 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rw6z\" (UniqueName: \"kubernetes.io/projected/e28b10f3-865f-4197-a095-f559444a1d2c-kube-api-access-9rw6z\") pod \"crc-debug-tkvbd\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.338560 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rw6z\" (UniqueName: \"kubernetes.io/projected/e28b10f3-865f-4197-a095-f559444a1d2c-kube-api-access-9rw6z\") pod \"crc-debug-tkvbd\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.338686 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e28b10f3-865f-4197-a095-f559444a1d2c-host\") pod \"crc-debug-tkvbd\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.338827 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e28b10f3-865f-4197-a095-f559444a1d2c-host\") pod \"crc-debug-tkvbd\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.378080 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rw6z\" (UniqueName: \"kubernetes.io/projected/e28b10f3-865f-4197-a095-f559444a1d2c-kube-api-access-9rw6z\") pod \"crc-debug-tkvbd\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: I1205 13:52:24.393491 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:52:24 crc kubenswrapper[4809]: W1205 13:52:24.441010 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode28b10f3_865f_4197_a095_f559444a1d2c.slice/crio-b0f36499a01bdc16b8db858db8ff5f01130cb004ad916777005c5afc28b92347 WatchSource:0}: Error finding container b0f36499a01bdc16b8db858db8ff5f01130cb004ad916777005c5afc28b92347: Status 404 returned error can't find the container with id b0f36499a01bdc16b8db858db8ff5f01130cb004ad916777005c5afc28b92347 Dec 05 13:52:25 crc kubenswrapper[4809]: I1205 13:52:25.141902 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" event={"ID":"e28b10f3-865f-4197-a095-f559444a1d2c","Type":"ContainerStarted","Data":"b0f36499a01bdc16b8db858db8ff5f01130cb004ad916777005c5afc28b92347"} Dec 05 13:52:39 crc kubenswrapper[4809]: I1205 13:52:39.302347 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" event={"ID":"e28b10f3-865f-4197-a095-f559444a1d2c","Type":"ContainerStarted","Data":"4415482fd90241bdc9f08a4ec855ae0b2f59148170ead9c5726d453c6935bc4a"} Dec 05 13:52:39 crc kubenswrapper[4809]: I1205 13:52:39.329834 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" podStartSLOduration=0.860850003 podStartE2EDuration="15.329814732s" podCreationTimestamp="2025-12-05 13:52:24 +0000 UTC" firstStartedPulling="2025-12-05 13:52:24.443533849 +0000 UTC m=+9839.834510407" lastFinishedPulling="2025-12-05 13:52:38.912498578 +0000 UTC m=+9854.303475136" observedRunningTime="2025-12-05 13:52:39.322383833 +0000 UTC m=+9854.713360401" watchObservedRunningTime="2025-12-05 13:52:39.329814732 +0000 UTC m=+9854.720791290" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.046285 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.046804 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.355146 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9thks"] Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.358056 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.367716 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9thks"] Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.483176 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q8vq\" (UniqueName: \"kubernetes.io/projected/dc7688b4-0499-4675-b5d7-a27e88aaec4b-kube-api-access-5q8vq\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.483327 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-utilities\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.483429 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-catalog-content\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.586115 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-catalog-content\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.586297 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q8vq\" (UniqueName: \"kubernetes.io/projected/dc7688b4-0499-4675-b5d7-a27e88aaec4b-kube-api-access-5q8vq\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.586372 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-utilities\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.586993 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-catalog-content\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.587045 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-utilities\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.605518 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q8vq\" (UniqueName: \"kubernetes.io/projected/dc7688b4-0499-4675-b5d7-a27e88aaec4b-kube-api-access-5q8vq\") pod \"certified-operators-9thks\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:44 crc kubenswrapper[4809]: I1205 13:52:44.688350 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:52:45 crc kubenswrapper[4809]: I1205 13:52:45.062897 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9thks"] Dec 05 13:52:45 crc kubenswrapper[4809]: I1205 13:52:45.365167 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9thks" event={"ID":"dc7688b4-0499-4675-b5d7-a27e88aaec4b","Type":"ContainerStarted","Data":"c24edbf5f3595a924265e35daf168a94723dc81f567d1cf0fc9fdcba3372d40e"} Dec 05 13:52:46 crc kubenswrapper[4809]: I1205 13:52:46.382952 4809 generic.go:334] "Generic (PLEG): container finished" podID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerID="d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e" exitCode=0 Dec 05 13:52:46 crc kubenswrapper[4809]: I1205 13:52:46.383022 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9thks" event={"ID":"dc7688b4-0499-4675-b5d7-a27e88aaec4b","Type":"ContainerDied","Data":"d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e"} Dec 05 13:52:54 crc kubenswrapper[4809]: I1205 13:52:54.775532 4809 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-kt4zh" podUID="6c091f2d-dbbe-4d80-af63-705da27c1dfe" containerName="ovnkube-controller" probeResult="failure" output="command timed out" Dec 05 13:52:57 crc kubenswrapper[4809]: I1205 13:52:57.341475 4809 trace.go:236] Trace[904242327]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-0" (05-Dec-2025 13:52:54.260) (total time: 3081ms): Dec 05 13:52:57 crc kubenswrapper[4809]: Trace[904242327]: [3.081056705s] [3.081056705s] END Dec 05 13:52:59 crc kubenswrapper[4809]: I1205 13:52:59.535455 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9thks" event={"ID":"dc7688b4-0499-4675-b5d7-a27e88aaec4b","Type":"ContainerStarted","Data":"04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0"} Dec 05 13:53:00 crc kubenswrapper[4809]: I1205 13:53:00.545562 4809 generic.go:334] "Generic (PLEG): container finished" podID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerID="04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0" exitCode=0 Dec 05 13:53:00 crc kubenswrapper[4809]: I1205 13:53:00.545753 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9thks" event={"ID":"dc7688b4-0499-4675-b5d7-a27e88aaec4b","Type":"ContainerDied","Data":"04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0"} Dec 05 13:53:01 crc kubenswrapper[4809]: I1205 13:53:01.559395 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9thks" event={"ID":"dc7688b4-0499-4675-b5d7-a27e88aaec4b","Type":"ContainerStarted","Data":"6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9"} Dec 05 13:53:01 crc kubenswrapper[4809]: I1205 13:53:01.581929 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9thks" podStartSLOduration=6.958525085 podStartE2EDuration="17.581911593s" podCreationTimestamp="2025-12-05 13:52:44 +0000 UTC" firstStartedPulling="2025-12-05 13:52:50.316152695 +0000 UTC m=+9865.707129253" lastFinishedPulling="2025-12-05 13:53:00.939539203 +0000 UTC m=+9876.330515761" observedRunningTime="2025-12-05 13:53:01.577473264 +0000 UTC m=+9876.968449862" watchObservedRunningTime="2025-12-05 13:53:01.581911593 +0000 UTC m=+9876.972888151" Dec 05 13:53:04 crc kubenswrapper[4809]: I1205 13:53:04.689366 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:53:04 crc kubenswrapper[4809]: I1205 13:53:04.689971 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:53:04 crc kubenswrapper[4809]: I1205 13:53:04.788683 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:53:08 crc kubenswrapper[4809]: I1205 13:53:08.624768 4809 generic.go:334] "Generic (PLEG): container finished" podID="e28b10f3-865f-4197-a095-f559444a1d2c" containerID="4415482fd90241bdc9f08a4ec855ae0b2f59148170ead9c5726d453c6935bc4a" exitCode=0 Dec 05 13:53:08 crc kubenswrapper[4809]: I1205 13:53:08.624849 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" event={"ID":"e28b10f3-865f-4197-a095-f559444a1d2c","Type":"ContainerDied","Data":"4415482fd90241bdc9f08a4ec855ae0b2f59148170ead9c5726d453c6935bc4a"} Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.290667 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.336098 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wpqvz/crc-debug-tkvbd"] Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.349246 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wpqvz/crc-debug-tkvbd"] Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.465562 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e28b10f3-865f-4197-a095-f559444a1d2c-host\") pod \"e28b10f3-865f-4197-a095-f559444a1d2c\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.465607 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rw6z\" (UniqueName: \"kubernetes.io/projected/e28b10f3-865f-4197-a095-f559444a1d2c-kube-api-access-9rw6z\") pod \"e28b10f3-865f-4197-a095-f559444a1d2c\" (UID: \"e28b10f3-865f-4197-a095-f559444a1d2c\") " Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.465691 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e28b10f3-865f-4197-a095-f559444a1d2c-host" (OuterVolumeSpecName: "host") pod "e28b10f3-865f-4197-a095-f559444a1d2c" (UID: "e28b10f3-865f-4197-a095-f559444a1d2c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.466190 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e28b10f3-865f-4197-a095-f559444a1d2c-host\") on node \"crc\" DevicePath \"\"" Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.488840 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e28b10f3-865f-4197-a095-f559444a1d2c-kube-api-access-9rw6z" (OuterVolumeSpecName: "kube-api-access-9rw6z") pod "e28b10f3-865f-4197-a095-f559444a1d2c" (UID: "e28b10f3-865f-4197-a095-f559444a1d2c"). InnerVolumeSpecName "kube-api-access-9rw6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.570173 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rw6z\" (UniqueName: \"kubernetes.io/projected/e28b10f3-865f-4197-a095-f559444a1d2c-kube-api-access-9rw6z\") on node \"crc\" DevicePath \"\"" Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.648945 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0f36499a01bdc16b8db858db8ff5f01130cb004ad916777005c5afc28b92347" Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.649025 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-tkvbd" Dec 05 13:53:10 crc kubenswrapper[4809]: I1205 13:53:10.886783 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e28b10f3-865f-4197-a095-f559444a1d2c" path="/var/lib/kubelet/pods/e28b10f3-865f-4197-a095-f559444a1d2c/volumes" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.586259 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wpqvz/crc-debug-bg4rs"] Dec 05 13:53:11 crc kubenswrapper[4809]: E1205 13:53:11.586918 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e28b10f3-865f-4197-a095-f559444a1d2c" containerName="container-00" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.586933 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="e28b10f3-865f-4197-a095-f559444a1d2c" containerName="container-00" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.587175 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="e28b10f3-865f-4197-a095-f559444a1d2c" containerName="container-00" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.587982 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.591696 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzlw4\" (UniqueName: \"kubernetes.io/projected/33233bd4-5c58-4a5c-bbd6-05854be07cc7-kube-api-access-fzlw4\") pod \"crc-debug-bg4rs\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.591847 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33233bd4-5c58-4a5c-bbd6-05854be07cc7-host\") pod \"crc-debug-bg4rs\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.693779 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzlw4\" (UniqueName: \"kubernetes.io/projected/33233bd4-5c58-4a5c-bbd6-05854be07cc7-kube-api-access-fzlw4\") pod \"crc-debug-bg4rs\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.694299 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33233bd4-5c58-4a5c-bbd6-05854be07cc7-host\") pod \"crc-debug-bg4rs\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.694464 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33233bd4-5c58-4a5c-bbd6-05854be07cc7-host\") pod \"crc-debug-bg4rs\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.723611 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzlw4\" (UniqueName: \"kubernetes.io/projected/33233bd4-5c58-4a5c-bbd6-05854be07cc7-kube-api-access-fzlw4\") pod \"crc-debug-bg4rs\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:11 crc kubenswrapper[4809]: I1205 13:53:11.905318 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:12 crc kubenswrapper[4809]: I1205 13:53:12.669000 4809 generic.go:334] "Generic (PLEG): container finished" podID="33233bd4-5c58-4a5c-bbd6-05854be07cc7" containerID="aeeb4573361c245684b8a56a8d9eee8a1532744753d18a5a9b1a57037f8fa5d2" exitCode=1 Dec 05 13:53:12 crc kubenswrapper[4809]: I1205 13:53:12.669199 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" event={"ID":"33233bd4-5c58-4a5c-bbd6-05854be07cc7","Type":"ContainerDied","Data":"aeeb4573361c245684b8a56a8d9eee8a1532744753d18a5a9b1a57037f8fa5d2"} Dec 05 13:53:12 crc kubenswrapper[4809]: I1205 13:53:12.670268 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" event={"ID":"33233bd4-5c58-4a5c-bbd6-05854be07cc7","Type":"ContainerStarted","Data":"2c9b1e4a4c6ea0f3630647514c122162997f973b10e3670da09187dbf7a6087d"} Dec 05 13:53:12 crc kubenswrapper[4809]: I1205 13:53:12.718055 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wpqvz/crc-debug-bg4rs"] Dec 05 13:53:12 crc kubenswrapper[4809]: I1205 13:53:12.728527 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wpqvz/crc-debug-bg4rs"] Dec 05 13:53:13 crc kubenswrapper[4809]: I1205 13:53:13.792880 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:13 crc kubenswrapper[4809]: I1205 13:53:13.937951 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33233bd4-5c58-4a5c-bbd6-05854be07cc7-host\") pod \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " Dec 05 13:53:13 crc kubenswrapper[4809]: I1205 13:53:13.938197 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzlw4\" (UniqueName: \"kubernetes.io/projected/33233bd4-5c58-4a5c-bbd6-05854be07cc7-kube-api-access-fzlw4\") pod \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\" (UID: \"33233bd4-5c58-4a5c-bbd6-05854be07cc7\") " Dec 05 13:53:13 crc kubenswrapper[4809]: I1205 13:53:13.938434 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/33233bd4-5c58-4a5c-bbd6-05854be07cc7-host" (OuterVolumeSpecName: "host") pod "33233bd4-5c58-4a5c-bbd6-05854be07cc7" (UID: "33233bd4-5c58-4a5c-bbd6-05854be07cc7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 13:53:13 crc kubenswrapper[4809]: I1205 13:53:13.939170 4809 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33233bd4-5c58-4a5c-bbd6-05854be07cc7-host\") on node \"crc\" DevicePath \"\"" Dec 05 13:53:13 crc kubenswrapper[4809]: I1205 13:53:13.944347 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33233bd4-5c58-4a5c-bbd6-05854be07cc7-kube-api-access-fzlw4" (OuterVolumeSpecName: "kube-api-access-fzlw4") pod "33233bd4-5c58-4a5c-bbd6-05854be07cc7" (UID: "33233bd4-5c58-4a5c-bbd6-05854be07cc7"). InnerVolumeSpecName "kube-api-access-fzlw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.040903 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzlw4\" (UniqueName: \"kubernetes.io/projected/33233bd4-5c58-4a5c-bbd6-05854be07cc7-kube-api-access-fzlw4\") on node \"crc\" DevicePath \"\"" Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.046970 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.047024 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.697115 4809 scope.go:117] "RemoveContainer" containerID="aeeb4573361c245684b8a56a8d9eee8a1532744753d18a5a9b1a57037f8fa5d2" Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.697598 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/crc-debug-bg4rs" Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.762379 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.850898 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9thks"] Dec 05 13:53:14 crc kubenswrapper[4809]: I1205 13:53:14.890075 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33233bd4-5c58-4a5c-bbd6-05854be07cc7" path="/var/lib/kubelet/pods/33233bd4-5c58-4a5c-bbd6-05854be07cc7/volumes" Dec 05 13:53:15 crc kubenswrapper[4809]: I1205 13:53:15.707898 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9thks" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="registry-server" containerID="cri-o://6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9" gracePeriod=2 Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.220300 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.287748 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q8vq\" (UniqueName: \"kubernetes.io/projected/dc7688b4-0499-4675-b5d7-a27e88aaec4b-kube-api-access-5q8vq\") pod \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.287934 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-catalog-content\") pod \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.287978 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-utilities\") pod \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\" (UID: \"dc7688b4-0499-4675-b5d7-a27e88aaec4b\") " Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.288782 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-utilities" (OuterVolumeSpecName: "utilities") pod "dc7688b4-0499-4675-b5d7-a27e88aaec4b" (UID: "dc7688b4-0499-4675-b5d7-a27e88aaec4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.297880 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc7688b4-0499-4675-b5d7-a27e88aaec4b-kube-api-access-5q8vq" (OuterVolumeSpecName: "kube-api-access-5q8vq") pod "dc7688b4-0499-4675-b5d7-a27e88aaec4b" (UID: "dc7688b4-0499-4675-b5d7-a27e88aaec4b"). InnerVolumeSpecName "kube-api-access-5q8vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.357168 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc7688b4-0499-4675-b5d7-a27e88aaec4b" (UID: "dc7688b4-0499-4675-b5d7-a27e88aaec4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.390674 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.390904 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q8vq\" (UniqueName: \"kubernetes.io/projected/dc7688b4-0499-4675-b5d7-a27e88aaec4b-kube-api-access-5q8vq\") on node \"crc\" DevicePath \"\"" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.390992 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7688b4-0499-4675-b5d7-a27e88aaec4b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.722818 4809 generic.go:334] "Generic (PLEG): container finished" podID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerID="6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9" exitCode=0 Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.722877 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9thks" event={"ID":"dc7688b4-0499-4675-b5d7-a27e88aaec4b","Type":"ContainerDied","Data":"6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9"} Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.722907 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9thks" event={"ID":"dc7688b4-0499-4675-b5d7-a27e88aaec4b","Type":"ContainerDied","Data":"c24edbf5f3595a924265e35daf168a94723dc81f567d1cf0fc9fdcba3372d40e"} Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.722925 4809 scope.go:117] "RemoveContainer" containerID="6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.722955 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9thks" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.754665 4809 scope.go:117] "RemoveContainer" containerID="04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.774860 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9thks"] Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.778990 4809 scope.go:117] "RemoveContainer" containerID="d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.791955 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9thks"] Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.831310 4809 scope.go:117] "RemoveContainer" containerID="6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9" Dec 05 13:53:16 crc kubenswrapper[4809]: E1205 13:53:16.831865 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9\": container with ID starting with 6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9 not found: ID does not exist" containerID="6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.831903 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9"} err="failed to get container status \"6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9\": rpc error: code = NotFound desc = could not find container \"6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9\": container with ID starting with 6e84eabb82c5d7132a32506f046ffd923c220862ee1b8990e45754065b49baf9 not found: ID does not exist" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.831924 4809 scope.go:117] "RemoveContainer" containerID="04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0" Dec 05 13:53:16 crc kubenswrapper[4809]: E1205 13:53:16.832427 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0\": container with ID starting with 04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0 not found: ID does not exist" containerID="04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.832469 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0"} err="failed to get container status \"04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0\": rpc error: code = NotFound desc = could not find container \"04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0\": container with ID starting with 04623d0bb13320d433c14a1c99fb7aa865e05f0fa7f0f1c693dbc1bc21e2e5e0 not found: ID does not exist" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.832486 4809 scope.go:117] "RemoveContainer" containerID="d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e" Dec 05 13:53:16 crc kubenswrapper[4809]: E1205 13:53:16.832904 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e\": container with ID starting with d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e not found: ID does not exist" containerID="d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.832943 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e"} err="failed to get container status \"d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e\": rpc error: code = NotFound desc = could not find container \"d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e\": container with ID starting with d7f18377526eacaa5c79df1b8cabfd44428d9fd60c0b85f3c906600cedf7b29e not found: ID does not exist" Dec 05 13:53:16 crc kubenswrapper[4809]: I1205 13:53:16.885035 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" path="/var/lib/kubelet/pods/dc7688b4-0499-4675-b5d7-a27e88aaec4b/volumes" Dec 05 13:53:44 crc kubenswrapper[4809]: I1205 13:53:44.046531 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 13:53:44 crc kubenswrapper[4809]: I1205 13:53:44.047221 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 13:53:44 crc kubenswrapper[4809]: I1205 13:53:44.047268 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 13:53:44 crc kubenswrapper[4809]: I1205 13:53:44.048051 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 13:53:44 crc kubenswrapper[4809]: I1205 13:53:44.048102 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" gracePeriod=600 Dec 05 13:53:44 crc kubenswrapper[4809]: E1205 13:53:44.173324 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:53:45 crc kubenswrapper[4809]: I1205 13:53:45.021009 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" exitCode=0 Dec 05 13:53:45 crc kubenswrapper[4809]: I1205 13:53:45.021081 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55"} Dec 05 13:53:45 crc kubenswrapper[4809]: I1205 13:53:45.021458 4809 scope.go:117] "RemoveContainer" containerID="ed4f444a78eb4a4541d40910e0e3d50538ccec421a579187e7645cf38d2ff4f0" Dec 05 13:53:45 crc kubenswrapper[4809]: I1205 13:53:45.022138 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:53:45 crc kubenswrapper[4809]: E1205 13:53:45.022534 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:53:55 crc kubenswrapper[4809]: I1205 13:53:55.873569 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:53:55 crc kubenswrapper[4809]: E1205 13:53:55.874558 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:54:07 crc kubenswrapper[4809]: I1205 13:54:07.873089 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:54:07 crc kubenswrapper[4809]: E1205 13:54:07.873857 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:54:21 crc kubenswrapper[4809]: I1205 13:54:21.871627 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:54:21 crc kubenswrapper[4809]: E1205 13:54:21.872343 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:54:35 crc kubenswrapper[4809]: I1205 13:54:35.872485 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:54:35 crc kubenswrapper[4809]: E1205 13:54:35.873367 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:54:49 crc kubenswrapper[4809]: I1205 13:54:49.872216 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:54:49 crc kubenswrapper[4809]: E1205 13:54:49.873289 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:55:04 crc kubenswrapper[4809]: I1205 13:55:04.881470 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:55:04 crc kubenswrapper[4809]: E1205 13:55:04.882959 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:55:15 crc kubenswrapper[4809]: I1205 13:55:15.873143 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:55:15 crc kubenswrapper[4809]: E1205 13:55:15.874259 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:55:29 crc kubenswrapper[4809]: I1205 13:55:29.872650 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:55:29 crc kubenswrapper[4809]: E1205 13:55:29.873567 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:55:43 crc kubenswrapper[4809]: I1205 13:55:43.872111 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:55:43 crc kubenswrapper[4809]: E1205 13:55:43.872921 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.816320 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mbql8"] Dec 05 13:55:53 crc kubenswrapper[4809]: E1205 13:55:53.817789 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="extract-content" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.817817 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="extract-content" Dec 05 13:55:53 crc kubenswrapper[4809]: E1205 13:55:53.817853 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33233bd4-5c58-4a5c-bbd6-05854be07cc7" containerName="container-00" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.817864 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="33233bd4-5c58-4a5c-bbd6-05854be07cc7" containerName="container-00" Dec 05 13:55:53 crc kubenswrapper[4809]: E1205 13:55:53.817885 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="extract-utilities" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.817897 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="extract-utilities" Dec 05 13:55:53 crc kubenswrapper[4809]: E1205 13:55:53.817926 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="registry-server" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.817936 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="registry-server" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.818304 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7688b4-0499-4675-b5d7-a27e88aaec4b" containerName="registry-server" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.818327 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="33233bd4-5c58-4a5c-bbd6-05854be07cc7" containerName="container-00" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.821621 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.834268 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbql8"] Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.873405 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-catalog-content\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.873888 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsnm6\" (UniqueName: \"kubernetes.io/projected/5ae918e3-375f-454c-b12e-14ecf2d83f90-kube-api-access-fsnm6\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.873916 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-utilities\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.976090 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-catalog-content\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.976449 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsnm6\" (UniqueName: \"kubernetes.io/projected/5ae918e3-375f-454c-b12e-14ecf2d83f90-kube-api-access-fsnm6\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.976502 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-utilities\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.976678 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-catalog-content\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:53 crc kubenswrapper[4809]: I1205 13:55:53.976979 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-utilities\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:54 crc kubenswrapper[4809]: I1205 13:55:54.004431 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsnm6\" (UniqueName: \"kubernetes.io/projected/5ae918e3-375f-454c-b12e-14ecf2d83f90-kube-api-access-fsnm6\") pod \"redhat-marketplace-mbql8\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:54 crc kubenswrapper[4809]: I1205 13:55:54.165215 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:55:54 crc kubenswrapper[4809]: I1205 13:55:54.711368 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbql8"] Dec 05 13:55:55 crc kubenswrapper[4809]: I1205 13:55:55.391932 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerID="d7778c852494c6e076bec04df9f661a036eaf2969a487d2ed424b380da87c48d" exitCode=0 Dec 05 13:55:55 crc kubenswrapper[4809]: I1205 13:55:55.392009 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbql8" event={"ID":"5ae918e3-375f-454c-b12e-14ecf2d83f90","Type":"ContainerDied","Data":"d7778c852494c6e076bec04df9f661a036eaf2969a487d2ed424b380da87c48d"} Dec 05 13:55:55 crc kubenswrapper[4809]: I1205 13:55:55.392281 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbql8" event={"ID":"5ae918e3-375f-454c-b12e-14ecf2d83f90","Type":"ContainerStarted","Data":"7e0a4630b80020c03a1c2bc8d86a180c2ee22a977c3ca1ca258bafbeeae5ba3d"} Dec 05 13:55:55 crc kubenswrapper[4809]: I1205 13:55:55.873441 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:55:55 crc kubenswrapper[4809]: E1205 13:55:55.873775 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:55:56 crc kubenswrapper[4809]: I1205 13:55:56.402208 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbql8" event={"ID":"5ae918e3-375f-454c-b12e-14ecf2d83f90","Type":"ContainerStarted","Data":"21c08d3f71f858803ddc386e163f1daa1e3ba637e3392d174cffd4a1d09146d1"} Dec 05 13:55:57 crc kubenswrapper[4809]: I1205 13:55:57.414653 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerID="21c08d3f71f858803ddc386e163f1daa1e3ba637e3392d174cffd4a1d09146d1" exitCode=0 Dec 05 13:55:57 crc kubenswrapper[4809]: I1205 13:55:57.414697 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbql8" event={"ID":"5ae918e3-375f-454c-b12e-14ecf2d83f90","Type":"ContainerDied","Data":"21c08d3f71f858803ddc386e163f1daa1e3ba637e3392d174cffd4a1d09146d1"} Dec 05 13:55:58 crc kubenswrapper[4809]: I1205 13:55:58.426153 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbql8" event={"ID":"5ae918e3-375f-454c-b12e-14ecf2d83f90","Type":"ContainerStarted","Data":"f99bea16192f1e83b5dc2f10fcd68ce6d0ec06150c3dbc1d36e75a236b0d71a2"} Dec 05 13:55:58 crc kubenswrapper[4809]: I1205 13:55:58.450836 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mbql8" podStartSLOduration=3.025377905 podStartE2EDuration="5.450812685s" podCreationTimestamp="2025-12-05 13:55:53 +0000 UTC" firstStartedPulling="2025-12-05 13:55:55.40032803 +0000 UTC m=+10050.791304588" lastFinishedPulling="2025-12-05 13:55:57.82576281 +0000 UTC m=+10053.216739368" observedRunningTime="2025-12-05 13:55:58.447105856 +0000 UTC m=+10053.838082424" watchObservedRunningTime="2025-12-05 13:55:58.450812685 +0000 UTC m=+10053.841789243" Dec 05 13:56:02 crc kubenswrapper[4809]: I1205 13:56:02.973874 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_374aeffe-f715-4f6e-af5b-4ff7f547d9c5/init-config-reloader/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.133325 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_374aeffe-f715-4f6e-af5b-4ff7f547d9c5/init-config-reloader/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.170808 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_374aeffe-f715-4f6e-af5b-4ff7f547d9c5/alertmanager/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.200159 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_374aeffe-f715-4f6e-af5b-4ff7f547d9c5/config-reloader/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.351237 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9768daa9-0e4c-4f29-bb4d-0c548a818691/aodh-api/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.437304 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9768daa9-0e4c-4f29-bb4d-0c548a818691/aodh-evaluator/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.539871 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9768daa9-0e4c-4f29-bb4d-0c548a818691/aodh-listener/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.581477 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9768daa9-0e4c-4f29-bb4d-0c548a818691/aodh-notifier/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.656529 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6454f6bb5d-n6n86_24422fbf-e081-4b3c-895c-91f5cfe2a340/barbican-api/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.764700 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6454f6bb5d-n6n86_24422fbf-e081-4b3c-895c-91f5cfe2a340/barbican-api-log/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.879494 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78f54888c6-ht5kn_db6726bf-d6e9-43fa-828f-c78798598723/barbican-keystone-listener/0.log" Dec 05 13:56:03 crc kubenswrapper[4809]: I1205 13:56:03.939222 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78f54888c6-ht5kn_db6726bf-d6e9-43fa-828f-c78798598723/barbican-keystone-listener-log/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.056573 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b9d486759-h8xx5_048fa7a0-bb2f-4e32-b3c3-0c5af443caf9/barbican-worker/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.154497 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b9d486759-h8xx5_048fa7a0-bb2f-4e32-b3c3-0c5af443caf9/barbican-worker-log/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.165979 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.166056 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.217703 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.245756 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-rjcr7_49d66a02-1ee1-4e87-b1f2-d7fbcd340eca/bootstrap-openstack-openstack-cell1/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.409738 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_63f6c3ad-0465-4149-98ae-da47e309692d/ceilometer-central-agent/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.477661 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_63f6c3ad-0465-4149-98ae-da47e309692d/ceilometer-notification-agent/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.530469 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_63f6c3ad-0465-4149-98ae-da47e309692d/proxy-httpd/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.544168 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.618094 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbql8"] Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.653598 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_63f6c3ad-0465-4149-98ae-da47e309692d/sg-core/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.761030 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-dn498_452f44ca-1d18-4f53-a39e-74b0b4d9c942/ceph-client-openstack-openstack-cell1/0.log" Dec 05 13:56:04 crc kubenswrapper[4809]: I1205 13:56:04.966431 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1d878828-def9-4192-adfc-1479f2e7a252/cinder-api/0.log" Dec 05 13:56:05 crc kubenswrapper[4809]: I1205 13:56:05.000754 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1d878828-def9-4192-adfc-1479f2e7a252/cinder-api-log/0.log" Dec 05 13:56:05 crc kubenswrapper[4809]: I1205 13:56:05.288469 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_e41e2028-117d-4e52-878b-0c722a1891aa/probe/0.log" Dec 05 13:56:05 crc kubenswrapper[4809]: I1205 13:56:05.304565 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b255ea1a-622c-4d9e-bf97-977a4f359ed5/cinder-scheduler/0.log" Dec 05 13:56:05 crc kubenswrapper[4809]: I1205 13:56:05.305574 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_e41e2028-117d-4e52-878b-0c722a1891aa/cinder-backup/0.log" Dec 05 13:56:05 crc kubenswrapper[4809]: I1205 13:56:05.531886 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b255ea1a-622c-4d9e-bf97-977a4f359ed5/probe/0.log" Dec 05 13:56:05 crc kubenswrapper[4809]: I1205 13:56:05.564900 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_2f939aa0-d8aa-4771-a4fe-326db3a407df/cinder-volume/0.log" Dec 05 13:56:06 crc kubenswrapper[4809]: I1205 13:56:06.271180 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-4r56d_0fbfa629-2f16-46c9-a67b-30b89444e688/configure-network-openstack-openstack-cell1/0.log" Dec 05 13:56:06 crc kubenswrapper[4809]: I1205 13:56:06.310675 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_2f939aa0-d8aa-4771-a4fe-326db3a407df/probe/0.log" Dec 05 13:56:06 crc kubenswrapper[4809]: I1205 13:56:06.504615 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mbql8" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="registry-server" containerID="cri-o://f99bea16192f1e83b5dc2f10fcd68ce6d0ec06150c3dbc1d36e75a236b0d71a2" gracePeriod=2 Dec 05 13:56:06 crc kubenswrapper[4809]: I1205 13:56:06.686011 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-bsvmn_cf82dcca-750d-458f-92ea-d4bc9b713f48/configure-os-openstack-openstack-cell1/0.log" Dec 05 13:56:06 crc kubenswrapper[4809]: I1205 13:56:06.693147 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58856dc695-gvqqw_07888d82-49b7-4107-ae3c-f34572b10fe0/init/0.log" Dec 05 13:56:06 crc kubenswrapper[4809]: I1205 13:56:06.890090 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58856dc695-gvqqw_07888d82-49b7-4107-ae3c-f34572b10fe0/dnsmasq-dns/0.log" Dec 05 13:56:06 crc kubenswrapper[4809]: I1205 13:56:06.934451 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58856dc695-gvqqw_07888d82-49b7-4107-ae3c-f34572b10fe0/init/0.log" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.045674 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-99s2k_2ba9f48f-f049-46da-b82b-62785eff9fd3/download-cache-openstack-openstack-cell1/0.log" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.166712 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_323c1d96-4a17-48b4-a300-57dd30a08cec/glance-httpd/0.log" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.188145 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_323c1d96-4a17-48b4-a300-57dd30a08cec/glance-log/0.log" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.280938 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0efb9902-75fb-4f9b-8966-8b70f61cd523/glance-httpd/0.log" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.315233 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0efb9902-75fb-4f9b-8966-8b70f61cd523/glance-log/0.log" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.531158 4809 generic.go:334] "Generic (PLEG): container finished" podID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerID="f99bea16192f1e83b5dc2f10fcd68ce6d0ec06150c3dbc1d36e75a236b0d71a2" exitCode=0 Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.531220 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbql8" event={"ID":"5ae918e3-375f-454c-b12e-14ecf2d83f90","Type":"ContainerDied","Data":"f99bea16192f1e83b5dc2f10fcd68ce6d0ec06150c3dbc1d36e75a236b0d71a2"} Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.531264 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mbql8" event={"ID":"5ae918e3-375f-454c-b12e-14ecf2d83f90","Type":"ContainerDied","Data":"7e0a4630b80020c03a1c2bc8d86a180c2ee22a977c3ca1ca258bafbeeae5ba3d"} Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.531278 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e0a4630b80020c03a1c2bc8d86a180c2ee22a977c3ca1ca258bafbeeae5ba3d" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.831879 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.903815 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-catalog-content\") pod \"5ae918e3-375f-454c-b12e-14ecf2d83f90\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.904323 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsnm6\" (UniqueName: \"kubernetes.io/projected/5ae918e3-375f-454c-b12e-14ecf2d83f90-kube-api-access-fsnm6\") pod \"5ae918e3-375f-454c-b12e-14ecf2d83f90\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.904407 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-utilities\") pod \"5ae918e3-375f-454c-b12e-14ecf2d83f90\" (UID: \"5ae918e3-375f-454c-b12e-14ecf2d83f90\") " Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.904927 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-utilities" (OuterVolumeSpecName: "utilities") pod "5ae918e3-375f-454c-b12e-14ecf2d83f90" (UID: "5ae918e3-375f-454c-b12e-14ecf2d83f90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.905103 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.912606 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae918e3-375f-454c-b12e-14ecf2d83f90-kube-api-access-fsnm6" (OuterVolumeSpecName: "kube-api-access-fsnm6") pod "5ae918e3-375f-454c-b12e-14ecf2d83f90" (UID: "5ae918e3-375f-454c-b12e-14ecf2d83f90"). InnerVolumeSpecName "kube-api-access-fsnm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:56:07 crc kubenswrapper[4809]: I1205 13:56:07.935120 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ae918e3-375f-454c-b12e-14ecf2d83f90" (UID: "5ae918e3-375f-454c-b12e-14ecf2d83f90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.006983 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsnm6\" (UniqueName: \"kubernetes.io/projected/5ae918e3-375f-454c-b12e-14ecf2d83f90-kube-api-access-fsnm6\") on node \"crc\" DevicePath \"\"" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.007028 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae918e3-375f-454c-b12e-14ecf2d83f90-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.118172 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-66d9fc74dc-mv9p7_6ebcec3a-3b2f-4b08-aa54-1701d35d272f/heat-api/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.208968 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-755878f84-9sw72_0b40e46c-2ed7-4277-8971-e366814ed23f/heat-cfnapi/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.308155 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-68cf595575-nttrk_39f3b59e-52c9-49a9-b63a-2129f74c2640/heat-engine/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.473582 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6db4cc6f8f-j562n_89b8338a-fd8b-44cd-8a96-9863586f40ac/horizon-log/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.523233 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6db4cc6f8f-j562n_89b8338a-fd8b-44cd-8a96-9863586f40ac/horizon/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.542430 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mbql8" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.548588 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-l9x7s_e7d9b2ef-dbdf-46f5-8618-49560290b2c4/install-certs-openstack-openstack-cell1/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.584160 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbql8"] Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.601712 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mbql8"] Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.742149 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-st5fn_8a9a80df-5202-4649-a667-f121649e4acc/install-os-openstack-openstack-cell1/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.873775 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:56:08 crc kubenswrapper[4809]: E1205 13:56:08.874200 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.882404 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29415661-mf6gb_8a1ae2c0-9d01-45d8-b5b4-536a0c76d271/keystone-cron/0.log" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.890550 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" path="/var/lib/kubelet/pods/5ae918e3-375f-454c-b12e-14ecf2d83f90/volumes" Dec 05 13:56:08 crc kubenswrapper[4809]: I1205 13:56:08.930154 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c849bd98c-xwxgc_7039dc11-ca87-46f6-ad62-503f8c4a4bbb/keystone-api/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.060255 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_810023b7-c0cb-4c82-9d79-387562b0b0d1/kube-state-metrics/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.165604 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-m8cjr_328ce8d0-f42c-4e66-9de5-0f0cac6a2994/libvirt-openstack-openstack-cell1/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.305504 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_fb7f996f-ece9-4cfc-9744-f331e75550e4/manila-api-log/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.337325 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_fb7f996f-ece9-4cfc-9744-f331e75550e4/manila-api/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.494843 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eec2e149-075e-4b45-9140-2eba7beef8c4/manila-scheduler/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.544369 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eec2e149-075e-4b45-9140-2eba7beef8c4/probe/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.595165 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_575f4d22-5809-4ca1-bbf7-fdbdbe305194/manila-share/0.log" Dec 05 13:56:09 crc kubenswrapper[4809]: I1205 13:56:09.656438 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_575f4d22-5809-4ca1-bbf7-fdbdbe305194/probe/0.log" Dec 05 13:56:10 crc kubenswrapper[4809]: I1205 13:56:10.026146 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cd8c99689-xmlvs_08b88df2-c726-4578-9f31-ccf6690e58a1/neutron-httpd/0.log" Dec 05 13:56:10 crc kubenswrapper[4809]: I1205 13:56:10.055836 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cd8c99689-xmlvs_08b88df2-c726-4578-9f31-ccf6690e58a1/neutron-api/0.log" Dec 05 13:56:10 crc kubenswrapper[4809]: I1205 13:56:10.125202 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-qd9cq_3b3d10d8-24e2-4f12-9007-3fd1368261c0/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 05 13:56:10 crc kubenswrapper[4809]: I1205 13:56:10.641199 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-2cr2r_31f3c035-3079-4cf5-ab18-b4a1db3fb3b7/neutron-metadata-openstack-openstack-cell1/0.log" Dec 05 13:56:10 crc kubenswrapper[4809]: I1205 13:56:10.760227 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-x8pdw_19eab448-bf57-4575-9a45-c50bf7b3fb77/neutron-sriov-openstack-openstack-cell1/0.log" Dec 05 13:56:11 crc kubenswrapper[4809]: I1205 13:56:11.024415 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_edda60ef-c084-4473-9ca8-c5f755a3248b/nova-api-api/0.log" Dec 05 13:56:11 crc kubenswrapper[4809]: I1205 13:56:11.177316 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_edda60ef-c084-4473-9ca8-c5f755a3248b/nova-api-log/0.log" Dec 05 13:56:11 crc kubenswrapper[4809]: I1205 13:56:11.243407 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b0e73347-b4a0-43ef-9a9c-022a54af26b7/nova-cell0-conductor-conductor/0.log" Dec 05 13:56:11 crc kubenswrapper[4809]: I1205 13:56:11.341458 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0ae38166-059b-4dd8-b4df-939f676eeaac/nova-cell1-conductor-conductor/0.log" Dec 05 13:56:11 crc kubenswrapper[4809]: I1205 13:56:11.622951 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_0af04389-9c23-4925-aa2f-c0466ae00de3/nova-cell1-novncproxy-novncproxy/0.log" Dec 05 13:56:11 crc kubenswrapper[4809]: I1205 13:56:11.727308 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9lmv_2959aa37-7f80-4b96-9284-c508595cce86/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.000028 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-gxdjl_d6fd01bc-fc42-443d-9b5b-cae7cff1f5b2/nova-cell1-openstack-openstack-cell1/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.105488 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_003a4f36-e5aa-482f-af8e-bd1297698e12/nova-metadata-log/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.147907 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_003a4f36-e5aa-482f-af8e-bd1297698e12/nova-metadata-metadata/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.372495 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59bccf98fd-m9c2j_6fd18514-6933-428e-95eb-de00e63a5acb/init/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.619470 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_3544290a-8953-442b-85af-1ae538e6115c/nova-scheduler-scheduler/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.643165 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59bccf98fd-m9c2j_6fd18514-6933-428e-95eb-de00e63a5acb/octavia-api-provider-agent/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.841728 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59bccf98fd-m9c2j_6fd18514-6933-428e-95eb-de00e63a5acb/init/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.883310 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-59bccf98fd-m9c2j_6fd18514-6933-428e-95eb-de00e63a5acb/octavia-api/0.log" Dec 05 13:56:12 crc kubenswrapper[4809]: I1205 13:56:12.927828 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-4zc9q_aff7629b-0b32-4e6a-999b-1a63d990acd9/init/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.118452 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-4zc9q_aff7629b-0b32-4e6a-999b-1a63d990acd9/init/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.178348 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-sf2cf_3e42fdbf-f030-4f02-8912-e957f00db980/init/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.339312 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-4zc9q_aff7629b-0b32-4e6a-999b-1a63d990acd9/octavia-healthmanager/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.431552 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-sf2cf_3e42fdbf-f030-4f02-8912-e957f00db980/init/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.474480 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-sf2cf_3e42fdbf-f030-4f02-8912-e957f00db980/octavia-housekeeping/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.581030 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-4gnqd_6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261/init/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.870009 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-4gnqd_6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261/octavia-amphora-httpd/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.877294 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-64hxq_d8b98d1c-896e-49cc-aead-f30b8afbf33c/init/0.log" Dec 05 13:56:13 crc kubenswrapper[4809]: I1205 13:56:13.909036 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-4gnqd_6bcbc9eb-9ab1-45d0-b7e3-b9ca40444261/init/0.log" Dec 05 13:56:14 crc kubenswrapper[4809]: I1205 13:56:14.135060 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-64hxq_d8b98d1c-896e-49cc-aead-f30b8afbf33c/init/0.log" Dec 05 13:56:14 crc kubenswrapper[4809]: I1205 13:56:14.202718 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-64hxq_d8b98d1c-896e-49cc-aead-f30b8afbf33c/octavia-rsyslog/0.log" Dec 05 13:56:14 crc kubenswrapper[4809]: I1205 13:56:14.206407 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-w28wb_e414c222-8d57-46b8-bd74-36f1e9206dcf/init/0.log" Dec 05 13:56:14 crc kubenswrapper[4809]: I1205 13:56:14.498313 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-w28wb_e414c222-8d57-46b8-bd74-36f1e9206dcf/init/0.log" Dec 05 13:56:14 crc kubenswrapper[4809]: I1205 13:56:14.563885 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_295ed14b-b2f8-42eb-a9c8-1fc24e74654e/mysql-bootstrap/0.log" Dec 05 13:56:14 crc kubenswrapper[4809]: I1205 13:56:14.574778 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-w28wb_e414c222-8d57-46b8-bd74-36f1e9206dcf/octavia-worker/0.log" Dec 05 13:56:15 crc kubenswrapper[4809]: I1205 13:56:15.847932 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_295ed14b-b2f8-42eb-a9c8-1fc24e74654e/galera/0.log" Dec 05 13:56:15 crc kubenswrapper[4809]: I1205 13:56:15.858324 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_295ed14b-b2f8-42eb-a9c8-1fc24e74654e/mysql-bootstrap/0.log" Dec 05 13:56:15 crc kubenswrapper[4809]: I1205 13:56:15.861945 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5976c836-3783-481b-8dcb-8b5725a16e4c/mysql-bootstrap/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.020389 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5976c836-3783-481b-8dcb-8b5725a16e4c/mysql-bootstrap/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.099280 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c49ded6e-3eb8-44f1-9125-5938485a644a/openstackclient/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.155092 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5976c836-3783-481b-8dcb-8b5725a16e4c/galera/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.333538 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-hm2jg_60ff38f8-1c69-42c3-8968-9276b9af9dce/ovn-controller/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.448883 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-khns6_40aadb59-e0c4-4370-9ce0-6899c84771ee/openstack-network-exporter/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.602081 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sfspc_273a55f4-4590-47d6-8428-69f365f8ee54/ovsdb-server-init/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.794551 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sfspc_273a55f4-4590-47d6-8428-69f365f8ee54/ovsdb-server-init/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.802515 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sfspc_273a55f4-4590-47d6-8428-69f365f8ee54/ovsdb-server/0.log" Dec 05 13:56:16 crc kubenswrapper[4809]: I1205 13:56:16.910145 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sfspc_273a55f4-4590-47d6-8428-69f365f8ee54/ovs-vswitchd/0.log" Dec 05 13:56:17 crc kubenswrapper[4809]: I1205 13:56:17.013174 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_24e1ca2d-52aa-48a6-95df-c0dd39dc83fa/ovn-northd/0.log" Dec 05 13:56:17 crc kubenswrapper[4809]: I1205 13:56:17.051590 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_24e1ca2d-52aa-48a6-95df-c0dd39dc83fa/openstack-network-exporter/0.log" Dec 05 13:56:17 crc kubenswrapper[4809]: I1205 13:56:17.744741 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9353c82b-1d9c-4516-984c-855b7aaf663e/openstack-network-exporter/0.log" Dec 05 13:56:17 crc kubenswrapper[4809]: I1205 13:56:17.813155 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-rvkzx_db861e9a-684e-4ca6-99b6-c07b45229909/ovn-openstack-openstack-cell1/0.log" Dec 05 13:56:17 crc kubenswrapper[4809]: I1205 13:56:17.939246 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9353c82b-1d9c-4516-984c-855b7aaf663e/ovsdbserver-nb/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.025255 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_24f0d020-eb48-422a-995d-072129b627d8/ovsdbserver-nb/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.042466 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_24f0d020-eb48-422a-995d-072129b627d8/openstack-network-exporter/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.286976 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_eadb4411-40eb-47a1-994d-f73bd3d4d4b2/openstack-network-exporter/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.301664 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_eadb4411-40eb-47a1-994d-f73bd3d4d4b2/ovsdbserver-nb/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.403194 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_30ce764d-74a0-46aa-b35f-c96bdaedca59/openstack-network-exporter/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.597760 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_30ce764d-74a0-46aa-b35f-c96bdaedca59/ovsdbserver-sb/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.642466 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_742477c9-f396-434f-bfd7-339df39432b8/openstack-network-exporter/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.708429 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_742477c9-f396-434f-bfd7-339df39432b8/ovsdbserver-sb/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.803871 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_95ec104d-e045-42f0-9b19-951db84ca0cb/openstack-network-exporter/0.log" Dec 05 13:56:18 crc kubenswrapper[4809]: I1205 13:56:18.973895 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_95ec104d-e045-42f0-9b19-951db84ca0cb/ovsdbserver-sb/0.log" Dec 05 13:56:19 crc kubenswrapper[4809]: I1205 13:56:19.147224 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84d8c96676-kw4cb_ffb727f4-12bb-40ea-b8ba-3bb33141727e/placement-api/0.log" Dec 05 13:56:19 crc kubenswrapper[4809]: I1205 13:56:19.237103 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84d8c96676-kw4cb_ffb727f4-12bb-40ea-b8ba-3bb33141727e/placement-log/0.log" Dec 05 13:56:19 crc kubenswrapper[4809]: I1205 13:56:19.266746 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cwmmxs_b8d0c6ee-25d0-4e37-aa15-da0204a934d3/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 05 13:56:19 crc kubenswrapper[4809]: I1205 13:56:19.454332 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0381cceb-e34d-4a34-bfb0-8673a52237dc/init-config-reloader/0.log" Dec 05 13:56:19 crc kubenswrapper[4809]: I1205 13:56:19.925412 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0381cceb-e34d-4a34-bfb0-8673a52237dc/config-reloader/0.log" Dec 05 13:56:19 crc kubenswrapper[4809]: I1205 13:56:19.935546 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0381cceb-e34d-4a34-bfb0-8673a52237dc/thanos-sidecar/0.log" Dec 05 13:56:19 crc kubenswrapper[4809]: I1205 13:56:19.976144 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0381cceb-e34d-4a34-bfb0-8673a52237dc/prometheus/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.010560 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0381cceb-e34d-4a34-bfb0-8673a52237dc/init-config-reloader/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.216299 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a621b075-5a72-4a2a-992f-a927ffb98a57/setup-container/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.265213 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_573f7284-f946-46f2-9a4d-55ab2520a344/memcached/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.396353 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a621b075-5a72-4a2a-992f-a927ffb98a57/rabbitmq/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.398898 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a621b075-5a72-4a2a-992f-a927ffb98a57/setup-container/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.448984 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5b994bc1-9a84-41c6-97df-6ee239d92968/setup-container/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.633381 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5b994bc1-9a84-41c6-97df-6ee239d92968/setup-container/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.740114 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-gqxn9_386049c0-b517-465a-a3b8-283127b5de78/reboot-os-openstack-openstack-cell1/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.804182 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5b994bc1-9a84-41c6-97df-6ee239d92968/rabbitmq/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.894146 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-fwxrn_43079fec-0ad2-472e-af20-6b102ed4284f/run-os-openstack-openstack-cell1/0.log" Dec 05 13:56:20 crc kubenswrapper[4809]: I1205 13:56:20.971905 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-nd5b4_bf168ba6-afe9-4d8e-b73f-1454c4ee6522/ssh-known-hosts-openstack/0.log" Dec 05 13:56:21 crc kubenswrapper[4809]: I1205 13:56:21.019269 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-7mb5p_0fb44e23-96d1-4a58-b7df-0522db74bb22/telemetry-openstack-openstack-cell1/0.log" Dec 05 13:56:21 crc kubenswrapper[4809]: I1205 13:56:21.235264 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-n4wl8_1e49a268-2565-4d37-aaa1-c256617a13a4/validate-network-openstack-openstack-cell1/0.log" Dec 05 13:56:21 crc kubenswrapper[4809]: I1205 13:56:21.241565 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-99vmj_3ee9bb21-400b-4e0a-966c-2625bcc76aae/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 05 13:56:21 crc kubenswrapper[4809]: I1205 13:56:21.872777 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:56:21 crc kubenswrapper[4809]: E1205 13:56:21.873215 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:56:33 crc kubenswrapper[4809]: I1205 13:56:33.872002 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:56:33 crc kubenswrapper[4809]: E1205 13:56:33.874395 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.014579 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-cc22d_f14d52a2-20e5-429a-8285-82ff53430a63/kube-rbac-proxy/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.201775 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-cc22d_f14d52a2-20e5-429a-8285-82ff53430a63/manager/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.272237 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-nmg6x_ad1e2739-6610-4b87-aec7-89ca78575409/kube-rbac-proxy/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.348432 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-nmg6x_ad1e2739-6610-4b87-aec7-89ca78575409/manager/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.454432 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-cjt5m_adae4108-5380-4358-b8bd-44ee6bb554af/kube-rbac-proxy/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.487310 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-cjt5m_adae4108-5380-4358-b8bd-44ee6bb554af/manager/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.674452 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb_3a3db807-0a5b-4fbb-ad41-86857e3ec884/util/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.819258 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb_3a3db807-0a5b-4fbb-ad41-86857e3ec884/pull/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.834990 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb_3a3db807-0a5b-4fbb-ad41-86857e3ec884/util/0.log" Dec 05 13:56:45 crc kubenswrapper[4809]: I1205 13:56:45.877085 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb_3a3db807-0a5b-4fbb-ad41-86857e3ec884/pull/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.010456 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb_3a3db807-0a5b-4fbb-ad41-86857e3ec884/util/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.035940 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb_3a3db807-0a5b-4fbb-ad41-86857e3ec884/extract/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.037776 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e48d19870fbb270f90e07dce4d9bad603b7d4900857291efd849a97a2e7glwb_3a3db807-0a5b-4fbb-ad41-86857e3ec884/pull/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.194972 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-plqqf_88757511-54a6-45b0-8bd8-90dddf1cae9f/kube-rbac-proxy/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.290988 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-9q4zp_9eec44c6-6e4d-4945-afef-ed95846e7b50/kube-rbac-proxy/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.379349 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-plqqf_88757511-54a6-45b0-8bd8-90dddf1cae9f/manager/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.444732 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-9q4zp_9eec44c6-6e4d-4945-afef-ed95846e7b50/manager/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.485416 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-57hcg_84c8e973-2683-4917-bea9-a7f44eadf397/kube-rbac-proxy/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.596409 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-57hcg_84c8e973-2683-4917-bea9-a7f44eadf397/manager/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.700649 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-k8cxw_6167f262-8efb-43c5-b065-4a2de04a183f/kube-rbac-proxy/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.852795 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-jgxpj_06bcaa51-2671-458b-9cd0-a4c73455b02a/kube-rbac-proxy/0.log" Dec 05 13:56:46 crc kubenswrapper[4809]: I1205 13:56:46.978356 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-k8cxw_6167f262-8efb-43c5-b065-4a2de04a183f/manager/0.log" Dec 05 13:56:47 crc kubenswrapper[4809]: I1205 13:56:47.177290 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-jgxpj_06bcaa51-2671-458b-9cd0-a4c73455b02a/manager/0.log" Dec 05 13:56:47 crc kubenswrapper[4809]: I1205 13:56:47.254089 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-jdbkf_e664f482-5587-4ad2-a35e-0fe439d6de39/kube-rbac-proxy/0.log" Dec 05 13:56:47 crc kubenswrapper[4809]: I1205 13:56:47.497453 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-jdbkf_e664f482-5587-4ad2-a35e-0fe439d6de39/manager/0.log" Dec 05 13:56:47 crc kubenswrapper[4809]: I1205 13:56:47.664401 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-57zmz_ab4b415e-554a-4e1b-a805-d628772bc6bc/kube-rbac-proxy/0.log" Dec 05 13:56:47 crc kubenswrapper[4809]: I1205 13:56:47.781745 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-57zmz_ab4b415e-554a-4e1b-a805-d628772bc6bc/manager/0.log" Dec 05 13:56:47 crc kubenswrapper[4809]: I1205 13:56:47.819941 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-7z7pv_2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15/kube-rbac-proxy/0.log" Dec 05 13:56:47 crc kubenswrapper[4809]: I1205 13:56:47.966153 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-7z7pv_2ed8d768-f5ea-4d08-9d06-cc6ad09c3e15/manager/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.086316 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-dmgm2_990aee67-4ef0-4a34-b228-d511085caf18/kube-rbac-proxy/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.181248 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-dmgm2_990aee67-4ef0-4a34-b228-d511085caf18/manager/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.325408 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-c27fb_094859bf-906c-4efa-b894-29db3182659b/kube-rbac-proxy/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.507315 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-c27fb_094859bf-906c-4efa-b894-29db3182659b/manager/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.510680 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-hx8w2_d47fc62a-29c5-4ebc-a3b4-45d6ce997907/kube-rbac-proxy/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.589527 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-hx8w2_d47fc62a-29c5-4ebc-a3b4-45d6ce997907/manager/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.711541 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc_7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83/kube-rbac-proxy/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.810685 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd46vrmc_7c5cc8f1-39d1-4a3a-bc1d-16e3d1175b83/manager/0.log" Dec 05 13:56:48 crc kubenswrapper[4809]: I1205 13:56:48.872079 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:56:48 crc kubenswrapper[4809]: E1205 13:56:48.872424 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:56:49 crc kubenswrapper[4809]: I1205 13:56:49.238179 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-md2wn_d8735bdd-709f-4c81-91a3-3ebca48c0d30/registry-server/0.log" Dec 05 13:56:49 crc kubenswrapper[4809]: I1205 13:56:49.374975 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6767b55986-hvln4_9d9c431e-1776-4dbc-b5be-5218c3117128/operator/0.log" Dec 05 13:56:49 crc kubenswrapper[4809]: I1205 13:56:49.559648 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mh2r6_53e22095-69df-44e2-aaea-7a6fbbbfce6c/kube-rbac-proxy/0.log" Dec 05 13:56:49 crc kubenswrapper[4809]: I1205 13:56:49.791915 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-5sz6t_8728b108-ea52-4024-bbba-d1985b5d9795/kube-rbac-proxy/0.log" Dec 05 13:56:49 crc kubenswrapper[4809]: I1205 13:56:49.882926 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mh2r6_53e22095-69df-44e2-aaea-7a6fbbbfce6c/manager/0.log" Dec 05 13:56:50 crc kubenswrapper[4809]: I1205 13:56:50.425286 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-5sz6t_8728b108-ea52-4024-bbba-d1985b5d9795/manager/0.log" Dec 05 13:56:50 crc kubenswrapper[4809]: I1205 13:56:50.472108 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-w8rhm_8bc5d720-4c3a-487b-a58d-4dc4c5760d6e/operator/0.log" Dec 05 13:56:50 crc kubenswrapper[4809]: I1205 13:56:50.688695 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-9m782_ac756b27-3bb1-4b8d-925c-15b41db189c8/kube-rbac-proxy/0.log" Dec 05 13:56:50 crc kubenswrapper[4809]: I1205 13:56:50.784375 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-ctckv_f586c1d3-4a51-48ff-be7f-d229d2770c10/kube-rbac-proxy/0.log" Dec 05 13:56:50 crc kubenswrapper[4809]: I1205 13:56:50.789155 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-9m782_ac756b27-3bb1-4b8d-925c-15b41db189c8/manager/0.log" Dec 05 13:56:51 crc kubenswrapper[4809]: I1205 13:56:51.004023 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-khmw4_adcbada3-1e83-4051-a036-f65cce0eb095/kube-rbac-proxy/0.log" Dec 05 13:56:51 crc kubenswrapper[4809]: I1205 13:56:51.205268 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-khmw4_adcbada3-1e83-4051-a036-f65cce0eb095/manager/0.log" Dec 05 13:56:51 crc kubenswrapper[4809]: I1205 13:56:51.470012 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64b69b8785-6nxr8_69cacfb4-8a61-4bd4-849f-94fc24d85453/manager/0.log" Dec 05 13:56:51 crc kubenswrapper[4809]: I1205 13:56:51.901059 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-vf5hg_a738af9f-fa2b-44ea-816a-135b351ef9d2/kube-rbac-proxy/0.log" Dec 05 13:56:51 crc kubenswrapper[4809]: I1205 13:56:51.912816 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-ctckv_f586c1d3-4a51-48ff-be7f-d229d2770c10/manager/0.log" Dec 05 13:56:51 crc kubenswrapper[4809]: I1205 13:56:51.984598 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-vf5hg_a738af9f-fa2b-44ea-816a-135b351ef9d2/manager/0.log" Dec 05 13:57:02 crc kubenswrapper[4809]: I1205 13:57:02.872023 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:57:02 crc kubenswrapper[4809]: E1205 13:57:02.872901 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:57:12 crc kubenswrapper[4809]: I1205 13:57:12.537371 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ml6sr_c166a15f-8b67-4236-8fd2-f3e84449c404/control-plane-machine-set-operator/0.log" Dec 05 13:57:12 crc kubenswrapper[4809]: I1205 13:57:12.737754 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6vmbb_9cf0c441-26db-4ac8-a1ed-4c2c90bc214c/kube-rbac-proxy/0.log" Dec 05 13:57:12 crc kubenswrapper[4809]: I1205 13:57:12.764608 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6vmbb_9cf0c441-26db-4ac8-a1ed-4c2c90bc214c/machine-api-operator/0.log" Dec 05 13:57:16 crc kubenswrapper[4809]: I1205 13:57:16.872270 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:57:16 crc kubenswrapper[4809]: E1205 13:57:16.872840 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:57:25 crc kubenswrapper[4809]: I1205 13:57:25.204343 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-qcfpj_32f2f5dc-9189-4cb7-9e85-e41acd2a9626/cert-manager-controller/0.log" Dec 05 13:57:25 crc kubenswrapper[4809]: I1205 13:57:25.324192 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-6n6jt_a46e32ba-4592-4ada-8289-d1329ae7e5df/cert-manager-cainjector/0.log" Dec 05 13:57:25 crc kubenswrapper[4809]: I1205 13:57:25.388826 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-2jdvh_10c849bc-151a-4e79-a288-ae9d082d6596/cert-manager-webhook/0.log" Dec 05 13:57:30 crc kubenswrapper[4809]: I1205 13:57:30.873846 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:57:30 crc kubenswrapper[4809]: E1205 13:57:30.875023 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:57:38 crc kubenswrapper[4809]: I1205 13:57:38.382033 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-cqk7j_701bb0f0-123a-4b4c-b302-c43597314c22/nmstate-console-plugin/0.log" Dec 05 13:57:38 crc kubenswrapper[4809]: I1205 13:57:38.545523 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4wrgc_f9d9930b-30f9-4e72-9ebe-f273ab71b519/kube-rbac-proxy/0.log" Dec 05 13:57:38 crc kubenswrapper[4809]: I1205 13:57:38.563587 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-8ql2q_94e36ccd-7287-4326-bf5e-864dd43c7701/nmstate-handler/0.log" Dec 05 13:57:38 crc kubenswrapper[4809]: I1205 13:57:38.685591 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4wrgc_f9d9930b-30f9-4e72-9ebe-f273ab71b519/nmstate-metrics/0.log" Dec 05 13:57:38 crc kubenswrapper[4809]: I1205 13:57:38.755473 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-xtkjh_bf3943ce-3694-4815-8d63-6f74dc7c3a9b/nmstate-operator/0.log" Dec 05 13:57:38 crc kubenswrapper[4809]: I1205 13:57:38.888847 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-pmtjt_f1b67fb0-2d57-4480-9fe5-8862894ce201/nmstate-webhook/0.log" Dec 05 13:57:41 crc kubenswrapper[4809]: I1205 13:57:41.872974 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:57:41 crc kubenswrapper[4809]: E1205 13:57:41.875136 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:57:54 crc kubenswrapper[4809]: I1205 13:57:54.536789 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hqh9p_aaa44bcc-a21a-45db-a639-c8336476e155/kube-rbac-proxy/0.log" Dec 05 13:57:54 crc kubenswrapper[4809]: I1205 13:57:54.880906 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-frr-files/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.015442 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-reloader/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.032647 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-frr-files/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.043274 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hqh9p_aaa44bcc-a21a-45db-a639-c8336476e155/controller/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.086971 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-metrics/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.203701 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-reloader/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.353962 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-metrics/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.363548 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-metrics/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.363800 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-frr-files/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.406235 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-reloader/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.534427 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-frr-files/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.550037 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-reloader/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.561190 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/cp-metrics/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.597462 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/controller/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.756783 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/kube-rbac-proxy/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.761858 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/frr-metrics/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.815197 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/kube-rbac-proxy-frr/0.log" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.872348 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:57:55 crc kubenswrapper[4809]: E1205 13:57:55.872715 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:57:55 crc kubenswrapper[4809]: I1205 13:57:55.967683 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/reloader/0.log" Dec 05 13:57:56 crc kubenswrapper[4809]: I1205 13:57:56.062684 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-89669_7f08f73d-94ad-41ad-8406-db20168bb7cd/frr-k8s-webhook-server/0.log" Dec 05 13:57:56 crc kubenswrapper[4809]: I1205 13:57:56.218468 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-567549d75d-v268p_7513fb12-ec94-4d79-ba5d-6dd2bad5c5b9/manager/0.log" Dec 05 13:57:56 crc kubenswrapper[4809]: I1205 13:57:56.397868 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86b59fcd6c-8p2pc_e3f8c150-501c-48f7-b111-44189a425a73/webhook-server/0.log" Dec 05 13:57:56 crc kubenswrapper[4809]: I1205 13:57:56.537852 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ns7rn_57d73721-c422-46ba-b347-b28c8027351a/kube-rbac-proxy/0.log" Dec 05 13:57:57 crc kubenswrapper[4809]: I1205 13:57:57.416278 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ns7rn_57d73721-c422-46ba-b347-b28c8027351a/speaker/0.log" Dec 05 13:57:59 crc kubenswrapper[4809]: I1205 13:57:59.076763 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgs44_864e3e5e-8812-4482-8cf0-a36c2e3c4d36/frr/0.log" Dec 05 13:58:06 crc kubenswrapper[4809]: I1205 13:58:06.872650 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:58:06 crc kubenswrapper[4809]: E1205 13:58:06.873486 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:58:11 crc kubenswrapper[4809]: I1205 13:58:11.688180 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j_2455a23f-14d3-41b8-9bef-e75ee783186a/util/0.log" Dec 05 13:58:11 crc kubenswrapper[4809]: I1205 13:58:11.899235 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j_2455a23f-14d3-41b8-9bef-e75ee783186a/util/0.log" Dec 05 13:58:11 crc kubenswrapper[4809]: I1205 13:58:11.906168 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j_2455a23f-14d3-41b8-9bef-e75ee783186a/pull/0.log" Dec 05 13:58:11 crc kubenswrapper[4809]: I1205 13:58:11.980920 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j_2455a23f-14d3-41b8-9bef-e75ee783186a/pull/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.132579 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j_2455a23f-14d3-41b8-9bef-e75ee783186a/pull/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.138286 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j_2455a23f-14d3-41b8-9bef-e75ee783186a/extract/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.152174 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931albz6j_2455a23f-14d3-41b8-9bef-e75ee783186a/util/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.311699 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf_cd91db76-778c-4208-83f3-9234765c7b8d/util/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.493998 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf_cd91db76-778c-4208-83f3-9234765c7b8d/pull/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.503525 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf_cd91db76-778c-4208-83f3-9234765c7b8d/util/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.504904 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf_cd91db76-778c-4208-83f3-9234765c7b8d/pull/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.647664 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf_cd91db76-778c-4208-83f3-9234765c7b8d/pull/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.652221 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf_cd91db76-778c-4208-83f3-9234765c7b8d/util/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.703150 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fhcsbf_cd91db76-778c-4208-83f3-9234765c7b8d/extract/0.log" Dec 05 13:58:12 crc kubenswrapper[4809]: I1205 13:58:12.803628 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd_59537950-4298-4f29-8513-d246da286b29/util/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.012500 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd_59537950-4298-4f29-8513-d246da286b29/pull/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.037973 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd_59537950-4298-4f29-8513-d246da286b29/util/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.044420 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd_59537950-4298-4f29-8513-d246da286b29/pull/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.201143 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd_59537950-4298-4f29-8513-d246da286b29/pull/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.207522 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd_59537950-4298-4f29-8513-d246da286b29/extract/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.226614 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210dk7pd_59537950-4298-4f29-8513-d246da286b29/util/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.372777 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt_b272d55e-85f0-4cb6-9bf4-360b5e2efc45/util/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.547602 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt_b272d55e-85f0-4cb6-9bf4-360b5e2efc45/pull/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.548502 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt_b272d55e-85f0-4cb6-9bf4-360b5e2efc45/pull/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.550686 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt_b272d55e-85f0-4cb6-9bf4-360b5e2efc45/util/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.725941 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt_b272d55e-85f0-4cb6-9bf4-360b5e2efc45/util/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.782194 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt_b272d55e-85f0-4cb6-9bf4-360b5e2efc45/extract/0.log" Dec 05 13:58:13 crc kubenswrapper[4809]: I1205 13:58:13.794185 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83b7ddt_b272d55e-85f0-4cb6-9bf4-360b5e2efc45/pull/0.log" Dec 05 13:58:14 crc kubenswrapper[4809]: I1205 13:58:14.701083 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nk4cv_a1b2282b-0f31-46c0-8b85-5e41e5bb1248/extract-utilities/0.log" Dec 05 13:58:14 crc kubenswrapper[4809]: I1205 13:58:14.903360 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nk4cv_a1b2282b-0f31-46c0-8b85-5e41e5bb1248/extract-utilities/0.log" Dec 05 13:58:14 crc kubenswrapper[4809]: I1205 13:58:14.933800 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nk4cv_a1b2282b-0f31-46c0-8b85-5e41e5bb1248/extract-content/0.log" Dec 05 13:58:14 crc kubenswrapper[4809]: I1205 13:58:14.941045 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nk4cv_a1b2282b-0f31-46c0-8b85-5e41e5bb1248/extract-content/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.059708 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nk4cv_a1b2282b-0f31-46c0-8b85-5e41e5bb1248/extract-utilities/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.111880 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nk4cv_a1b2282b-0f31-46c0-8b85-5e41e5bb1248/extract-content/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.212861 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wkvnz_b07cc015-c350-4282-84e3-7bd761097773/extract-utilities/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.375124 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wkvnz_b07cc015-c350-4282-84e3-7bd761097773/extract-utilities/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.388366 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wkvnz_b07cc015-c350-4282-84e3-7bd761097773/extract-content/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.401586 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wkvnz_b07cc015-c350-4282-84e3-7bd761097773/extract-content/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.625857 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wkvnz_b07cc015-c350-4282-84e3-7bd761097773/extract-utilities/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.631257 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wkvnz_b07cc015-c350-4282-84e3-7bd761097773/extract-content/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.870675 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vpwkx_b892319b-0c09-48a4-85af-dae860c7c31e/extract-utilities/0.log" Dec 05 13:58:15 crc kubenswrapper[4809]: I1205 13:58:15.927998 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-t4mbz_dc5dc5f8-b0d8-4b61-b19b-d2970ef6f1cd/marketplace-operator/0.log" Dec 05 13:58:16 crc kubenswrapper[4809]: I1205 13:58:16.669137 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vpwkx_b892319b-0c09-48a4-85af-dae860c7c31e/extract-content/0.log" Dec 05 13:58:16 crc kubenswrapper[4809]: I1205 13:58:16.676410 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vpwkx_b892319b-0c09-48a4-85af-dae860c7c31e/extract-utilities/0.log" Dec 05 13:58:16 crc kubenswrapper[4809]: I1205 13:58:16.713465 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vpwkx_b892319b-0c09-48a4-85af-dae860c7c31e/extract-content/0.log" Dec 05 13:58:16 crc kubenswrapper[4809]: I1205 13:58:16.978674 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vpwkx_b892319b-0c09-48a4-85af-dae860c7c31e/extract-utilities/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.014154 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vpwkx_b892319b-0c09-48a4-85af-dae860c7c31e/extract-content/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.224467 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pm6h_724c514f-9629-482f-aa1d-f750ed364603/extract-utilities/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.428759 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pm6h_724c514f-9629-482f-aa1d-f750ed364603/extract-utilities/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.462337 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pm6h_724c514f-9629-482f-aa1d-f750ed364603/extract-content/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.486046 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vpwkx_b892319b-0c09-48a4-85af-dae860c7c31e/registry-server/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.594949 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nk4cv_a1b2282b-0f31-46c0-8b85-5e41e5bb1248/registry-server/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.666338 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pm6h_724c514f-9629-482f-aa1d-f750ed364603/extract-content/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.765261 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wkvnz_b07cc015-c350-4282-84e3-7bd761097773/registry-server/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.852730 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pm6h_724c514f-9629-482f-aa1d-f750ed364603/extract-content/0.log" Dec 05 13:58:17 crc kubenswrapper[4809]: I1205 13:58:17.893708 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pm6h_724c514f-9629-482f-aa1d-f750ed364603/extract-utilities/0.log" Dec 05 13:58:19 crc kubenswrapper[4809]: I1205 13:58:19.318144 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4pm6h_724c514f-9629-482f-aa1d-f750ed364603/registry-server/0.log" Dec 05 13:58:21 crc kubenswrapper[4809]: I1205 13:58:21.872677 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:58:21 crc kubenswrapper[4809]: E1205 13:58:21.873530 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:58:30 crc kubenswrapper[4809]: I1205 13:58:30.726848 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-srt4m_2412d41f-89ce-4153-9c3a-c281e9ed6406/prometheus-operator/0.log" Dec 05 13:58:30 crc kubenswrapper[4809]: I1205 13:58:30.894991 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-66cf5c6d97-7ghsf_a5bf61ed-e17c-47f9-9d7e-d13bf3bef1ea/prometheus-operator-admission-webhook/0.log" Dec 05 13:58:30 crc kubenswrapper[4809]: I1205 13:58:30.962286 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-66cf5c6d97-xdvwx_60278de2-7102-42b7-b617-567279d19923/prometheus-operator-admission-webhook/0.log" Dec 05 13:58:31 crc kubenswrapper[4809]: I1205 13:58:31.106959 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-5zfk8_5e48c383-a2d1-4a9c-a17f-1d6b157ef997/operator/0.log" Dec 05 13:58:31 crc kubenswrapper[4809]: I1205 13:58:31.142247 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-hqwhc_bd2d4a74-3f06-4ad0-8387-a6fe2e8f2d63/perses-operator/0.log" Dec 05 13:58:36 crc kubenswrapper[4809]: I1205 13:58:36.871869 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:58:36 crc kubenswrapper[4809]: E1205 13:58:36.872756 4809 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s4hls_openshift-machine-config-operator(87517ccc-4f19-430e-9d03-d71da4cbe196)\"" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" Dec 05 13:58:48 crc kubenswrapper[4809]: I1205 13:58:48.393190 4809 scope.go:117] "RemoveContainer" containerID="4415482fd90241bdc9f08a4ec855ae0b2f59148170ead9c5726d453c6935bc4a" Dec 05 13:58:50 crc kubenswrapper[4809]: I1205 13:58:50.872829 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 13:58:51 crc kubenswrapper[4809]: I1205 13:58:51.254235 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"20d6b0d37271ab8f5e15d5e928773abe0709694fd1b9c70bdac63e02ace2508e"} Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.011018 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ns2zb"] Dec 05 13:59:16 crc kubenswrapper[4809]: E1205 13:59:16.014559 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="registry-server" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.014589 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="registry-server" Dec 05 13:59:16 crc kubenswrapper[4809]: E1205 13:59:16.014622 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="extract-utilities" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.014658 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="extract-utilities" Dec 05 13:59:16 crc kubenswrapper[4809]: E1205 13:59:16.014710 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="extract-content" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.014722 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="extract-content" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.015092 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae918e3-375f-454c-b12e-14ecf2d83f90" containerName="registry-server" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.018140 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.037148 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ns2zb"] Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.073730 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp44l\" (UniqueName: \"kubernetes.io/projected/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-kube-api-access-sp44l\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.073882 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-utilities\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.073914 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-catalog-content\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.176330 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-utilities\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.176387 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-catalog-content\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.176477 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp44l\" (UniqueName: \"kubernetes.io/projected/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-kube-api-access-sp44l\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.177260 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-utilities\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.177476 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-catalog-content\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.196115 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp44l\" (UniqueName: \"kubernetes.io/projected/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-kube-api-access-sp44l\") pod \"redhat-operators-ns2zb\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.346769 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:16 crc kubenswrapper[4809]: I1205 13:59:16.843955 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ns2zb"] Dec 05 13:59:17 crc kubenswrapper[4809]: I1205 13:59:17.557077 4809 generic.go:334] "Generic (PLEG): container finished" podID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerID="4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d" exitCode=0 Dec 05 13:59:17 crc kubenswrapper[4809]: I1205 13:59:17.557204 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ns2zb" event={"ID":"8b4c9636-9848-4f49-bd3e-3ec3ecefff51","Type":"ContainerDied","Data":"4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d"} Dec 05 13:59:17 crc kubenswrapper[4809]: I1205 13:59:17.557624 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ns2zb" event={"ID":"8b4c9636-9848-4f49-bd3e-3ec3ecefff51","Type":"ContainerStarted","Data":"61d91c96e99fb15826b14ed8415b2c2b2173a57f97d629ea7df562f793f73352"} Dec 05 13:59:17 crc kubenswrapper[4809]: I1205 13:59:17.566232 4809 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 13:59:18 crc kubenswrapper[4809]: I1205 13:59:18.568404 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ns2zb" event={"ID":"8b4c9636-9848-4f49-bd3e-3ec3ecefff51","Type":"ContainerStarted","Data":"3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1"} Dec 05 13:59:21 crc kubenswrapper[4809]: I1205 13:59:21.618778 4809 generic.go:334] "Generic (PLEG): container finished" podID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerID="3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1" exitCode=0 Dec 05 13:59:21 crc kubenswrapper[4809]: I1205 13:59:21.618914 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ns2zb" event={"ID":"8b4c9636-9848-4f49-bd3e-3ec3ecefff51","Type":"ContainerDied","Data":"3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1"} Dec 05 13:59:22 crc kubenswrapper[4809]: I1205 13:59:22.642492 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ns2zb" event={"ID":"8b4c9636-9848-4f49-bd3e-3ec3ecefff51","Type":"ContainerStarted","Data":"3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369"} Dec 05 13:59:22 crc kubenswrapper[4809]: I1205 13:59:22.664668 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ns2zb" podStartSLOduration=3.202319171 podStartE2EDuration="7.664627696s" podCreationTimestamp="2025-12-05 13:59:15 +0000 UTC" firstStartedPulling="2025-12-05 13:59:17.565915961 +0000 UTC m=+10252.956892529" lastFinishedPulling="2025-12-05 13:59:22.028224476 +0000 UTC m=+10257.419201054" observedRunningTime="2025-12-05 13:59:22.661230465 +0000 UTC m=+10258.052207033" watchObservedRunningTime="2025-12-05 13:59:22.664627696 +0000 UTC m=+10258.055604254" Dec 05 13:59:26 crc kubenswrapper[4809]: I1205 13:59:26.347936 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:26 crc kubenswrapper[4809]: I1205 13:59:26.349182 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:27 crc kubenswrapper[4809]: I1205 13:59:27.417557 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ns2zb" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="registry-server" probeResult="failure" output=< Dec 05 13:59:27 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 13:59:27 crc kubenswrapper[4809]: > Dec 05 13:59:36 crc kubenswrapper[4809]: I1205 13:59:36.412986 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:36 crc kubenswrapper[4809]: I1205 13:59:36.473860 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:36 crc kubenswrapper[4809]: I1205 13:59:36.660687 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ns2zb"] Dec 05 13:59:37 crc kubenswrapper[4809]: I1205 13:59:37.835224 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ns2zb" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="registry-server" containerID="cri-o://3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369" gracePeriod=2 Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.361094 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.512869 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-utilities\") pod \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.512943 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp44l\" (UniqueName: \"kubernetes.io/projected/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-kube-api-access-sp44l\") pod \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.513037 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-catalog-content\") pod \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\" (UID: \"8b4c9636-9848-4f49-bd3e-3ec3ecefff51\") " Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.513548 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-utilities" (OuterVolumeSpecName: "utilities") pod "8b4c9636-9848-4f49-bd3e-3ec3ecefff51" (UID: "8b4c9636-9848-4f49-bd3e-3ec3ecefff51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.513762 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.520832 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-kube-api-access-sp44l" (OuterVolumeSpecName: "kube-api-access-sp44l") pod "8b4c9636-9848-4f49-bd3e-3ec3ecefff51" (UID: "8b4c9636-9848-4f49-bd3e-3ec3ecefff51"). InnerVolumeSpecName "kube-api-access-sp44l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.618000 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp44l\" (UniqueName: \"kubernetes.io/projected/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-kube-api-access-sp44l\") on node \"crc\" DevicePath \"\"" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.699913 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b4c9636-9848-4f49-bd3e-3ec3ecefff51" (UID: "8b4c9636-9848-4f49-bd3e-3ec3ecefff51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.720002 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b4c9636-9848-4f49-bd3e-3ec3ecefff51-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.848227 4809 generic.go:334] "Generic (PLEG): container finished" podID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerID="3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369" exitCode=0 Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.848279 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ns2zb" event={"ID":"8b4c9636-9848-4f49-bd3e-3ec3ecefff51","Type":"ContainerDied","Data":"3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369"} Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.848310 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ns2zb" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.848343 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ns2zb" event={"ID":"8b4c9636-9848-4f49-bd3e-3ec3ecefff51","Type":"ContainerDied","Data":"61d91c96e99fb15826b14ed8415b2c2b2173a57f97d629ea7df562f793f73352"} Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.848369 4809 scope.go:117] "RemoveContainer" containerID="3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.889357 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ns2zb"] Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.893752 4809 scope.go:117] "RemoveContainer" containerID="3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.895383 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ns2zb"] Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.925009 4809 scope.go:117] "RemoveContainer" containerID="4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.975164 4809 scope.go:117] "RemoveContainer" containerID="3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369" Dec 05 13:59:38 crc kubenswrapper[4809]: E1205 13:59:38.975868 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369\": container with ID starting with 3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369 not found: ID does not exist" containerID="3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.975923 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369"} err="failed to get container status \"3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369\": rpc error: code = NotFound desc = could not find container \"3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369\": container with ID starting with 3fa68bb37c8e58e6d1fd33c34fe4ede2d20d6aa89664d66287f39ba97dea1369 not found: ID does not exist" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.975956 4809 scope.go:117] "RemoveContainer" containerID="3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1" Dec 05 13:59:38 crc kubenswrapper[4809]: E1205 13:59:38.976827 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1\": container with ID starting with 3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1 not found: ID does not exist" containerID="3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.976880 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1"} err="failed to get container status \"3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1\": rpc error: code = NotFound desc = could not find container \"3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1\": container with ID starting with 3cc56e86a6f61d6408a1c94071c13179d311edb1ac636a2a7617480936976aa1 not found: ID does not exist" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.976913 4809 scope.go:117] "RemoveContainer" containerID="4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d" Dec 05 13:59:38 crc kubenswrapper[4809]: E1205 13:59:38.981645 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d\": container with ID starting with 4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d not found: ID does not exist" containerID="4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d" Dec 05 13:59:38 crc kubenswrapper[4809]: I1205 13:59:38.981679 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d"} err="failed to get container status \"4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d\": rpc error: code = NotFound desc = could not find container \"4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d\": container with ID starting with 4b9d541a0caf8f355a044bc8d43df425f16f9f5730d70347977e299c1af5652d not found: ID does not exist" Dec 05 13:59:40 crc kubenswrapper[4809]: I1205 13:59:40.901414 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" path="/var/lib/kubelet/pods/8b4c9636-9848-4f49-bd3e-3ec3ecefff51/volumes" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.730587 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-khntv"] Dec 05 13:59:59 crc kubenswrapper[4809]: E1205 13:59:59.731524 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="extract-content" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.731536 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="extract-content" Dec 05 13:59:59 crc kubenswrapper[4809]: E1205 13:59:59.731564 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="extract-utilities" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.731571 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="extract-utilities" Dec 05 13:59:59 crc kubenswrapper[4809]: E1205 13:59:59.731594 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="registry-server" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.731600 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="registry-server" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.731823 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b4c9636-9848-4f49-bd3e-3ec3ecefff51" containerName="registry-server" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.733505 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.748345 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-khntv"] Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.775180 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g7vb\" (UniqueName: \"kubernetes.io/projected/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-kube-api-access-2g7vb\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.775259 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-utilities\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.775287 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-catalog-content\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.877427 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-utilities\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.877491 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-catalog-content\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.878216 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-catalog-content\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.878669 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-utilities\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.878775 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g7vb\" (UniqueName: \"kubernetes.io/projected/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-kube-api-access-2g7vb\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 13:59:59 crc kubenswrapper[4809]: I1205 13:59:59.905724 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g7vb\" (UniqueName: \"kubernetes.io/projected/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-kube-api-access-2g7vb\") pod \"community-operators-khntv\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.063201 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.196590 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9"] Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.198758 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.201475 4809 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.201798 4809 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.215377 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9"] Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.298945 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d5d186-da86-4d63-a2e8-1216b6d74120-config-volume\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.298990 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d5d186-da86-4d63-a2e8-1216b6d74120-secret-volume\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.299075 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7fls\" (UniqueName: \"kubernetes.io/projected/b1d5d186-da86-4d63-a2e8-1216b6d74120-kube-api-access-z7fls\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.400980 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d5d186-da86-4d63-a2e8-1216b6d74120-config-volume\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.401029 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d5d186-da86-4d63-a2e8-1216b6d74120-secret-volume\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.401086 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7fls\" (UniqueName: \"kubernetes.io/projected/b1d5d186-da86-4d63-a2e8-1216b6d74120-kube-api-access-z7fls\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.402295 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d5d186-da86-4d63-a2e8-1216b6d74120-config-volume\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.408012 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d5d186-da86-4d63-a2e8-1216b6d74120-secret-volume\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.425680 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7fls\" (UniqueName: \"kubernetes.io/projected/b1d5d186-da86-4d63-a2e8-1216b6d74120-kube-api-access-z7fls\") pod \"collect-profiles-29415720-4l2m9\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.560098 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:00 crc kubenswrapper[4809]: W1205 14:00:00.653551 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4c14b92_c709_4c01_89ff_f6d8fb75dadb.slice/crio-2736ea225de0d57cf3bfff903698a39f5f55cfea3802ee67d1122f2996a54717 WatchSource:0}: Error finding container 2736ea225de0d57cf3bfff903698a39f5f55cfea3802ee67d1122f2996a54717: Status 404 returned error can't find the container with id 2736ea225de0d57cf3bfff903698a39f5f55cfea3802ee67d1122f2996a54717 Dec 05 14:00:00 crc kubenswrapper[4809]: I1205 14:00:00.657148 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-khntv"] Dec 05 14:00:01 crc kubenswrapper[4809]: I1205 14:00:01.077965 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9"] Dec 05 14:00:01 crc kubenswrapper[4809]: W1205 14:00:01.088541 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1d5d186_da86_4d63_a2e8_1216b6d74120.slice/crio-7f5ae93ea336b52b44b75ef3eb051bb417c7e8516081728e2aa3a9ed9dca0ad4 WatchSource:0}: Error finding container 7f5ae93ea336b52b44b75ef3eb051bb417c7e8516081728e2aa3a9ed9dca0ad4: Status 404 returned error can't find the container with id 7f5ae93ea336b52b44b75ef3eb051bb417c7e8516081728e2aa3a9ed9dca0ad4 Dec 05 14:00:01 crc kubenswrapper[4809]: I1205 14:00:01.183728 4809 generic.go:334] "Generic (PLEG): container finished" podID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerID="33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c" exitCode=0 Dec 05 14:00:01 crc kubenswrapper[4809]: I1205 14:00:01.183794 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khntv" event={"ID":"c4c14b92-c709-4c01-89ff-f6d8fb75dadb","Type":"ContainerDied","Data":"33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c"} Dec 05 14:00:01 crc kubenswrapper[4809]: I1205 14:00:01.183820 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khntv" event={"ID":"c4c14b92-c709-4c01-89ff-f6d8fb75dadb","Type":"ContainerStarted","Data":"2736ea225de0d57cf3bfff903698a39f5f55cfea3802ee67d1122f2996a54717"} Dec 05 14:00:01 crc kubenswrapper[4809]: I1205 14:00:01.185605 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" event={"ID":"b1d5d186-da86-4d63-a2e8-1216b6d74120","Type":"ContainerStarted","Data":"7f5ae93ea336b52b44b75ef3eb051bb417c7e8516081728e2aa3a9ed9dca0ad4"} Dec 05 14:00:01 crc kubenswrapper[4809]: E1205 14:00:01.653163 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1d5d186_da86_4d63_a2e8_1216b6d74120.slice/crio-82252dc8156a2f1a8c447d5432a3d5b3c5245816b9ef00d7306cfa558c1b3646.scope\": RecentStats: unable to find data in memory cache]" Dec 05 14:00:02 crc kubenswrapper[4809]: I1205 14:00:02.196304 4809 generic.go:334] "Generic (PLEG): container finished" podID="b1d5d186-da86-4d63-a2e8-1216b6d74120" containerID="82252dc8156a2f1a8c447d5432a3d5b3c5245816b9ef00d7306cfa558c1b3646" exitCode=0 Dec 05 14:00:02 crc kubenswrapper[4809]: I1205 14:00:02.196395 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" event={"ID":"b1d5d186-da86-4d63-a2e8-1216b6d74120","Type":"ContainerDied","Data":"82252dc8156a2f1a8c447d5432a3d5b3c5245816b9ef00d7306cfa558c1b3646"} Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.602591 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.696834 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d5d186-da86-4d63-a2e8-1216b6d74120-secret-volume\") pod \"b1d5d186-da86-4d63-a2e8-1216b6d74120\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.697027 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d5d186-da86-4d63-a2e8-1216b6d74120-config-volume\") pod \"b1d5d186-da86-4d63-a2e8-1216b6d74120\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.697177 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7fls\" (UniqueName: \"kubernetes.io/projected/b1d5d186-da86-4d63-a2e8-1216b6d74120-kube-api-access-z7fls\") pod \"b1d5d186-da86-4d63-a2e8-1216b6d74120\" (UID: \"b1d5d186-da86-4d63-a2e8-1216b6d74120\") " Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.710984 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d5d186-da86-4d63-a2e8-1216b6d74120-config-volume" (OuterVolumeSpecName: "config-volume") pod "b1d5d186-da86-4d63-a2e8-1216b6d74120" (UID: "b1d5d186-da86-4d63-a2e8-1216b6d74120"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.731048 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d5d186-da86-4d63-a2e8-1216b6d74120-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b1d5d186-da86-4d63-a2e8-1216b6d74120" (UID: "b1d5d186-da86-4d63-a2e8-1216b6d74120"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.735737 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d5d186-da86-4d63-a2e8-1216b6d74120-kube-api-access-z7fls" (OuterVolumeSpecName: "kube-api-access-z7fls") pod "b1d5d186-da86-4d63-a2e8-1216b6d74120" (UID: "b1d5d186-da86-4d63-a2e8-1216b6d74120"). InnerVolumeSpecName "kube-api-access-z7fls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.800049 4809 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d5d186-da86-4d63-a2e8-1216b6d74120-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.800274 4809 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d5d186-da86-4d63-a2e8-1216b6d74120-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:03 crc kubenswrapper[4809]: I1205 14:00:03.800285 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7fls\" (UniqueName: \"kubernetes.io/projected/b1d5d186-da86-4d63-a2e8-1216b6d74120-kube-api-access-z7fls\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:04 crc kubenswrapper[4809]: I1205 14:00:04.219297 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" event={"ID":"b1d5d186-da86-4d63-a2e8-1216b6d74120","Type":"ContainerDied","Data":"7f5ae93ea336b52b44b75ef3eb051bb417c7e8516081728e2aa3a9ed9dca0ad4"} Dec 05 14:00:04 crc kubenswrapper[4809]: I1205 14:00:04.219341 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f5ae93ea336b52b44b75ef3eb051bb417c7e8516081728e2aa3a9ed9dca0ad4" Dec 05 14:00:04 crc kubenswrapper[4809]: I1205 14:00:04.219403 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415720-4l2m9" Dec 05 14:00:04 crc kubenswrapper[4809]: I1205 14:00:04.682620 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn"] Dec 05 14:00:04 crc kubenswrapper[4809]: I1205 14:00:04.693235 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415675-qzxkn"] Dec 05 14:00:04 crc kubenswrapper[4809]: I1205 14:00:04.899728 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59fcdcce-9ede-4874-912d-c22ed395fa2d" path="/var/lib/kubelet/pods/59fcdcce-9ede-4874-912d-c22ed395fa2d/volumes" Dec 05 14:00:06 crc kubenswrapper[4809]: I1205 14:00:06.245512 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khntv" event={"ID":"c4c14b92-c709-4c01-89ff-f6d8fb75dadb","Type":"ContainerStarted","Data":"ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3"} Dec 05 14:00:07 crc kubenswrapper[4809]: I1205 14:00:07.255137 4809 generic.go:334] "Generic (PLEG): container finished" podID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerID="ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3" exitCode=0 Dec 05 14:00:07 crc kubenswrapper[4809]: I1205 14:00:07.255176 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khntv" event={"ID":"c4c14b92-c709-4c01-89ff-f6d8fb75dadb","Type":"ContainerDied","Data":"ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3"} Dec 05 14:00:09 crc kubenswrapper[4809]: I1205 14:00:09.279292 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khntv" event={"ID":"c4c14b92-c709-4c01-89ff-f6d8fb75dadb","Type":"ContainerStarted","Data":"a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31"} Dec 05 14:00:09 crc kubenswrapper[4809]: I1205 14:00:09.315320 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-khntv" podStartSLOduration=3.859477755 podStartE2EDuration="10.315300444s" podCreationTimestamp="2025-12-05 13:59:59 +0000 UTC" firstStartedPulling="2025-12-05 14:00:01.18714598 +0000 UTC m=+10296.578122538" lastFinishedPulling="2025-12-05 14:00:07.642968669 +0000 UTC m=+10303.033945227" observedRunningTime="2025-12-05 14:00:09.298669848 +0000 UTC m=+10304.689646426" watchObservedRunningTime="2025-12-05 14:00:09.315300444 +0000 UTC m=+10304.706276992" Dec 05 14:00:10 crc kubenswrapper[4809]: I1205 14:00:10.064799 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:10 crc kubenswrapper[4809]: I1205 14:00:10.065330 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:11 crc kubenswrapper[4809]: I1205 14:00:11.115612 4809 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-khntv" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="registry-server" probeResult="failure" output=< Dec 05 14:00:11 crc kubenswrapper[4809]: timeout: failed to connect service ":50051" within 1s Dec 05 14:00:11 crc kubenswrapper[4809]: > Dec 05 14:00:20 crc kubenswrapper[4809]: I1205 14:00:20.717227 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:20 crc kubenswrapper[4809]: I1205 14:00:20.791258 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:20 crc kubenswrapper[4809]: I1205 14:00:20.955843 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-khntv"] Dec 05 14:00:22 crc kubenswrapper[4809]: I1205 14:00:22.469934 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-khntv" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="registry-server" containerID="cri-o://a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31" gracePeriod=2 Dec 05 14:00:22 crc kubenswrapper[4809]: I1205 14:00:22.987302 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.050054 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-utilities\") pod \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.050103 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g7vb\" (UniqueName: \"kubernetes.io/projected/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-kube-api-access-2g7vb\") pod \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.050204 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-catalog-content\") pod \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\" (UID: \"c4c14b92-c709-4c01-89ff-f6d8fb75dadb\") " Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.050903 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-utilities" (OuterVolumeSpecName: "utilities") pod "c4c14b92-c709-4c01-89ff-f6d8fb75dadb" (UID: "c4c14b92-c709-4c01-89ff-f6d8fb75dadb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.055854 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-kube-api-access-2g7vb" (OuterVolumeSpecName: "kube-api-access-2g7vb") pod "c4c14b92-c709-4c01-89ff-f6d8fb75dadb" (UID: "c4c14b92-c709-4c01-89ff-f6d8fb75dadb"). InnerVolumeSpecName "kube-api-access-2g7vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.107578 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4c14b92-c709-4c01-89ff-f6d8fb75dadb" (UID: "c4c14b92-c709-4c01-89ff-f6d8fb75dadb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.152765 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.152807 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g7vb\" (UniqueName: \"kubernetes.io/projected/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-kube-api-access-2g7vb\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.152824 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c14b92-c709-4c01-89ff-f6d8fb75dadb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.482600 4809 generic.go:334] "Generic (PLEG): container finished" podID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerID="a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31" exitCode=0 Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.482658 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khntv" event={"ID":"c4c14b92-c709-4c01-89ff-f6d8fb75dadb","Type":"ContainerDied","Data":"a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31"} Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.482890 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khntv" event={"ID":"c4c14b92-c709-4c01-89ff-f6d8fb75dadb","Type":"ContainerDied","Data":"2736ea225de0d57cf3bfff903698a39f5f55cfea3802ee67d1122f2996a54717"} Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.482909 4809 scope.go:117] "RemoveContainer" containerID="a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.482697 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khntv" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.519040 4809 scope.go:117] "RemoveContainer" containerID="ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.529589 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-khntv"] Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.547362 4809 scope.go:117] "RemoveContainer" containerID="33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.554131 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-khntv"] Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.598923 4809 scope.go:117] "RemoveContainer" containerID="a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31" Dec 05 14:00:23 crc kubenswrapper[4809]: E1205 14:00:23.599418 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31\": container with ID starting with a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31 not found: ID does not exist" containerID="a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.599456 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31"} err="failed to get container status \"a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31\": rpc error: code = NotFound desc = could not find container \"a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31\": container with ID starting with a60db975c54494a07aa26617fafe22001bd3db68dd309eeb4ddd5a7c8a8c3f31 not found: ID does not exist" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.599487 4809 scope.go:117] "RemoveContainer" containerID="ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3" Dec 05 14:00:23 crc kubenswrapper[4809]: E1205 14:00:23.599860 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3\": container with ID starting with ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3 not found: ID does not exist" containerID="ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.599900 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3"} err="failed to get container status \"ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3\": rpc error: code = NotFound desc = could not find container \"ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3\": container with ID starting with ce4ba3b4fbac3f6a654b8a33986bed0b9847346c00496ee0ce4843e4c58480c3 not found: ID does not exist" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.599928 4809 scope.go:117] "RemoveContainer" containerID="33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c" Dec 05 14:00:23 crc kubenswrapper[4809]: E1205 14:00:23.600294 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c\": container with ID starting with 33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c not found: ID does not exist" containerID="33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c" Dec 05 14:00:23 crc kubenswrapper[4809]: I1205 14:00:23.600339 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c"} err="failed to get container status \"33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c\": rpc error: code = NotFound desc = could not find container \"33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c\": container with ID starting with 33506e2478c3f4c8d801e0434ba81f118617ed5ccb6081565ac4664bc378589c not found: ID does not exist" Dec 05 14:00:24 crc kubenswrapper[4809]: I1205 14:00:24.893527 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" path="/var/lib/kubelet/pods/c4c14b92-c709-4c01-89ff-f6d8fb75dadb/volumes" Dec 05 14:00:46 crc kubenswrapper[4809]: I1205 14:00:46.791066 4809 generic.go:334] "Generic (PLEG): container finished" podID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerID="a6a64210cf5c442ee8a44a35ff8bda9f12a6f87911ee8eb4650b647bbc0cb0d7" exitCode=0 Dec 05 14:00:46 crc kubenswrapper[4809]: I1205 14:00:46.791199 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" event={"ID":"b7e68161-aadf-4fcd-bfe0-f254200d4e5b","Type":"ContainerDied","Data":"a6a64210cf5c442ee8a44a35ff8bda9f12a6f87911ee8eb4650b647bbc0cb0d7"} Dec 05 14:00:46 crc kubenswrapper[4809]: I1205 14:00:46.792445 4809 scope.go:117] "RemoveContainer" containerID="a6a64210cf5c442ee8a44a35ff8bda9f12a6f87911ee8eb4650b647bbc0cb0d7" Dec 05 14:00:47 crc kubenswrapper[4809]: I1205 14:00:47.079704 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wpqvz_must-gather-pc8r6_b7e68161-aadf-4fcd-bfe0-f254200d4e5b/gather/0.log" Dec 05 14:00:49 crc kubenswrapper[4809]: I1205 14:00:49.024117 4809 scope.go:117] "RemoveContainer" containerID="6d55b7fd746429610367ce5d8a919c4e5fb87868396e626cf584a6af4996839d" Dec 05 14:00:56 crc kubenswrapper[4809]: I1205 14:00:56.081120 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wpqvz/must-gather-pc8r6"] Dec 05 14:00:56 crc kubenswrapper[4809]: I1205 14:00:56.081922 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerName="copy" containerID="cri-o://de2804e07062b4cd8939735fc30028607b439b7ef61115aedc2e63e7f9bbbebd" gracePeriod=2 Dec 05 14:00:56 crc kubenswrapper[4809]: I1205 14:00:56.094800 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wpqvz/must-gather-pc8r6"] Dec 05 14:00:56 crc kubenswrapper[4809]: I1205 14:00:56.932303 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wpqvz_must-gather-pc8r6_b7e68161-aadf-4fcd-bfe0-f254200d4e5b/copy/0.log" Dec 05 14:00:56 crc kubenswrapper[4809]: I1205 14:00:56.933172 4809 generic.go:334] "Generic (PLEG): container finished" podID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerID="de2804e07062b4cd8939735fc30028607b439b7ef61115aedc2e63e7f9bbbebd" exitCode=143 Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.282107 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wpqvz_must-gather-pc8r6_b7e68161-aadf-4fcd-bfe0-f254200d4e5b/copy/0.log" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.282778 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.448239 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d74qw\" (UniqueName: \"kubernetes.io/projected/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-kube-api-access-d74qw\") pod \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.448736 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-must-gather-output\") pod \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\" (UID: \"b7e68161-aadf-4fcd-bfe0-f254200d4e5b\") " Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.455724 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-kube-api-access-d74qw" (OuterVolumeSpecName: "kube-api-access-d74qw") pod "b7e68161-aadf-4fcd-bfe0-f254200d4e5b" (UID: "b7e68161-aadf-4fcd-bfe0-f254200d4e5b"). InnerVolumeSpecName "kube-api-access-d74qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.552038 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d74qw\" (UniqueName: \"kubernetes.io/projected/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-kube-api-access-d74qw\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.687690 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b7e68161-aadf-4fcd-bfe0-f254200d4e5b" (UID: "b7e68161-aadf-4fcd-bfe0-f254200d4e5b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.756132 4809 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b7e68161-aadf-4fcd-bfe0-f254200d4e5b-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.945339 4809 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wpqvz_must-gather-pc8r6_b7e68161-aadf-4fcd-bfe0-f254200d4e5b/copy/0.log" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.945794 4809 scope.go:117] "RemoveContainer" containerID="de2804e07062b4cd8939735fc30028607b439b7ef61115aedc2e63e7f9bbbebd" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.945865 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wpqvz/must-gather-pc8r6" Dec 05 14:00:57 crc kubenswrapper[4809]: I1205 14:00:57.969172 4809 scope.go:117] "RemoveContainer" containerID="a6a64210cf5c442ee8a44a35ff8bda9f12a6f87911ee8eb4650b647bbc0cb0d7" Dec 05 14:00:58 crc kubenswrapper[4809]: I1205 14:00:58.883437 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" path="/var/lib/kubelet/pods/b7e68161-aadf-4fcd-bfe0-f254200d4e5b/volumes" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.172376 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29415721-692n5"] Dec 05 14:01:00 crc kubenswrapper[4809]: E1205 14:01:00.173446 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerName="copy" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.173468 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerName="copy" Dec 05 14:01:00 crc kubenswrapper[4809]: E1205 14:01:00.173491 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d5d186-da86-4d63-a2e8-1216b6d74120" containerName="collect-profiles" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.173502 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d5d186-da86-4d63-a2e8-1216b6d74120" containerName="collect-profiles" Dec 05 14:01:00 crc kubenswrapper[4809]: E1205 14:01:00.173534 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerName="gather" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.173545 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerName="gather" Dec 05 14:01:00 crc kubenswrapper[4809]: E1205 14:01:00.173567 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="registry-server" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.173577 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="registry-server" Dec 05 14:01:00 crc kubenswrapper[4809]: E1205 14:01:00.173598 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="extract-content" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.173608 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="extract-content" Dec 05 14:01:00 crc kubenswrapper[4809]: E1205 14:01:00.173677 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="extract-utilities" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.173688 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="extract-utilities" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.173986 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d5d186-da86-4d63-a2e8-1216b6d74120" containerName="collect-profiles" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.174030 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4c14b92-c709-4c01-89ff-f6d8fb75dadb" containerName="registry-server" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.174052 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerName="copy" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.174067 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e68161-aadf-4fcd-bfe0-f254200d4e5b" containerName="gather" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.175277 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.183574 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415721-692n5"] Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.314483 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-combined-ca-bundle\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.314928 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbswn\" (UniqueName: \"kubernetes.io/projected/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-kube-api-access-jbswn\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.314993 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-config-data\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.315091 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-fernet-keys\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.417610 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbswn\" (UniqueName: \"kubernetes.io/projected/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-kube-api-access-jbswn\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.417719 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-config-data\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.417790 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-fernet-keys\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.417816 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-combined-ca-bundle\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.423608 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-combined-ca-bundle\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.423660 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-config-data\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.423867 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-fernet-keys\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.435737 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbswn\" (UniqueName: \"kubernetes.io/projected/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-kube-api-access-jbswn\") pod \"keystone-cron-29415721-692n5\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:00 crc kubenswrapper[4809]: I1205 14:01:00.528437 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:01 crc kubenswrapper[4809]: I1205 14:01:01.016198 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415721-692n5"] Dec 05 14:01:01 crc kubenswrapper[4809]: I1205 14:01:01.990124 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415721-692n5" event={"ID":"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66","Type":"ContainerStarted","Data":"033b1c9ac4f69388e8dd4fa01403bd2f882692ae5fe711d53b186e9758c55785"} Dec 05 14:01:01 crc kubenswrapper[4809]: I1205 14:01:01.990510 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415721-692n5" event={"ID":"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66","Type":"ContainerStarted","Data":"c15ee272b8d4089b47cdbb6318f8e0ef7e3fe618d7474a7dbf69da42758be188"} Dec 05 14:01:02 crc kubenswrapper[4809]: I1205 14:01:02.015625 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29415721-692n5" podStartSLOduration=2.0156043 podStartE2EDuration="2.0156043s" podCreationTimestamp="2025-12-05 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 14:01:02.008363176 +0000 UTC m=+10357.399339764" watchObservedRunningTime="2025-12-05 14:01:02.0156043 +0000 UTC m=+10357.406580848" Dec 05 14:01:04 crc kubenswrapper[4809]: I1205 14:01:04.008663 4809 generic.go:334] "Generic (PLEG): container finished" podID="bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" containerID="033b1c9ac4f69388e8dd4fa01403bd2f882692ae5fe711d53b186e9758c55785" exitCode=0 Dec 05 14:01:04 crc kubenswrapper[4809]: I1205 14:01:04.008712 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415721-692n5" event={"ID":"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66","Type":"ContainerDied","Data":"033b1c9ac4f69388e8dd4fa01403bd2f882692ae5fe711d53b186e9758c55785"} Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.450690 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.647680 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbswn\" (UniqueName: \"kubernetes.io/projected/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-kube-api-access-jbswn\") pod \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.647787 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-fernet-keys\") pod \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.647867 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-combined-ca-bundle\") pod \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.647921 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-config-data\") pod \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\" (UID: \"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66\") " Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.654343 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" (UID: "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.665652 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-kube-api-access-jbswn" (OuterVolumeSpecName: "kube-api-access-jbswn") pod "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" (UID: "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66"). InnerVolumeSpecName "kube-api-access-jbswn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.683869 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" (UID: "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.714278 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-config-data" (OuterVolumeSpecName: "config-data") pod "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" (UID: "bc8f773c-19b0-4fd4-8b47-22e4e7ceef66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.749888 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbswn\" (UniqueName: \"kubernetes.io/projected/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-kube-api-access-jbswn\") on node \"crc\" DevicePath \"\"" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.749922 4809 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.749933 4809 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 14:01:05 crc kubenswrapper[4809]: I1205 14:01:05.749941 4809 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc8f773c-19b0-4fd4-8b47-22e4e7ceef66-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 14:01:06 crc kubenswrapper[4809]: I1205 14:01:06.034245 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415721-692n5" event={"ID":"bc8f773c-19b0-4fd4-8b47-22e4e7ceef66","Type":"ContainerDied","Data":"c15ee272b8d4089b47cdbb6318f8e0ef7e3fe618d7474a7dbf69da42758be188"} Dec 05 14:01:06 crc kubenswrapper[4809]: I1205 14:01:06.034293 4809 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c15ee272b8d4089b47cdbb6318f8e0ef7e3fe618d7474a7dbf69da42758be188" Dec 05 14:01:06 crc kubenswrapper[4809]: I1205 14:01:06.034377 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415721-692n5" Dec 05 14:01:14 crc kubenswrapper[4809]: I1205 14:01:14.046582 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 14:01:14 crc kubenswrapper[4809]: I1205 14:01:14.047178 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 14:01:44 crc kubenswrapper[4809]: I1205 14:01:44.046503 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 14:01:44 crc kubenswrapper[4809]: I1205 14:01:44.047178 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 14:02:14 crc kubenswrapper[4809]: I1205 14:02:14.047278 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 14:02:14 crc kubenswrapper[4809]: I1205 14:02:14.048196 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 14:02:14 crc kubenswrapper[4809]: I1205 14:02:14.048314 4809 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" Dec 05 14:02:14 crc kubenswrapper[4809]: I1205 14:02:14.851553 4809 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20d6b0d37271ab8f5e15d5e928773abe0709694fd1b9c70bdac63e02ace2508e"} pod="openshift-machine-config-operator/machine-config-daemon-s4hls" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 14:02:14 crc kubenswrapper[4809]: I1205 14:02:14.851693 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" containerID="cri-o://20d6b0d37271ab8f5e15d5e928773abe0709694fd1b9c70bdac63e02ace2508e" gracePeriod=600 Dec 05 14:02:15 crc kubenswrapper[4809]: I1205 14:02:15.864977 4809 generic.go:334] "Generic (PLEG): container finished" podID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerID="20d6b0d37271ab8f5e15d5e928773abe0709694fd1b9c70bdac63e02ace2508e" exitCode=0 Dec 05 14:02:15 crc kubenswrapper[4809]: I1205 14:02:15.865053 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerDied","Data":"20d6b0d37271ab8f5e15d5e928773abe0709694fd1b9c70bdac63e02ace2508e"} Dec 05 14:02:15 crc kubenswrapper[4809]: I1205 14:02:15.865569 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" event={"ID":"87517ccc-4f19-430e-9d03-d71da4cbe196","Type":"ContainerStarted","Data":"d578777a9ec6299f16c9b838299f30fabd7474932792b8ae8ed45eebca6a5b19"} Dec 05 14:02:15 crc kubenswrapper[4809]: I1205 14:02:15.865594 4809 scope.go:117] "RemoveContainer" containerID="ade2380f1b0b26e777f217c05526415fb25756ed5495dedfa4b18772039ead55" Dec 05 14:02:49 crc kubenswrapper[4809]: I1205 14:02:49.145066 4809 scope.go:117] "RemoveContainer" containerID="d7778c852494c6e076bec04df9f661a036eaf2969a487d2ed424b380da87c48d" Dec 05 14:02:49 crc kubenswrapper[4809]: I1205 14:02:49.174290 4809 scope.go:117] "RemoveContainer" containerID="f99bea16192f1e83b5dc2f10fcd68ce6d0ec06150c3dbc1d36e75a236b0d71a2" Dec 05 14:02:49 crc kubenswrapper[4809]: I1205 14:02:49.262079 4809 scope.go:117] "RemoveContainer" containerID="21c08d3f71f858803ddc386e163f1daa1e3ba637e3392d174cffd4a1d09146d1" Dec 05 14:03:33 crc kubenswrapper[4809]: I1205 14:03:33.822858 4809 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4snt6"] Dec 05 14:03:33 crc kubenswrapper[4809]: E1205 14:03:33.824082 4809 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" containerName="keystone-cron" Dec 05 14:03:33 crc kubenswrapper[4809]: I1205 14:03:33.824100 4809 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" containerName="keystone-cron" Dec 05 14:03:33 crc kubenswrapper[4809]: I1205 14:03:33.824378 4809 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc8f773c-19b0-4fd4-8b47-22e4e7ceef66" containerName="keystone-cron" Dec 05 14:03:33 crc kubenswrapper[4809]: I1205 14:03:33.826357 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:33 crc kubenswrapper[4809]: I1205 14:03:33.844174 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4snt6"] Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.021932 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtr8t\" (UniqueName: \"kubernetes.io/projected/b5cd96f2-7836-4939-b188-a22420a964e6-kube-api-access-xtr8t\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.022533 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-catalog-content\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.022951 4809 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-utilities\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.125012 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-utilities\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.125127 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtr8t\" (UniqueName: \"kubernetes.io/projected/b5cd96f2-7836-4939-b188-a22420a964e6-kube-api-access-xtr8t\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.125265 4809 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-catalog-content\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.125935 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-catalog-content\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.126121 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-utilities\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.154476 4809 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtr8t\" (UniqueName: \"kubernetes.io/projected/b5cd96f2-7836-4939-b188-a22420a964e6-kube-api-access-xtr8t\") pod \"certified-operators-4snt6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.157457 4809 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.725153 4809 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4snt6"] Dec 05 14:03:34 crc kubenswrapper[4809]: W1205 14:03:34.732012 4809 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5cd96f2_7836_4939_b188_a22420a964e6.slice/crio-f8203264c201007546502c60875c5307775cd41541cc700c3e81ed25d013c2b2 WatchSource:0}: Error finding container f8203264c201007546502c60875c5307775cd41541cc700c3e81ed25d013c2b2: Status 404 returned error can't find the container with id f8203264c201007546502c60875c5307775cd41541cc700c3e81ed25d013c2b2 Dec 05 14:03:34 crc kubenswrapper[4809]: I1205 14:03:34.849790 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4snt6" event={"ID":"b5cd96f2-7836-4939-b188-a22420a964e6","Type":"ContainerStarted","Data":"f8203264c201007546502c60875c5307775cd41541cc700c3e81ed25d013c2b2"} Dec 05 14:03:44 crc kubenswrapper[4809]: I1205 14:03:44.952722 4809 generic.go:334] "Generic (PLEG): container finished" podID="b5cd96f2-7836-4939-b188-a22420a964e6" containerID="c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb" exitCode=0 Dec 05 14:03:44 crc kubenswrapper[4809]: I1205 14:03:44.952967 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4snt6" event={"ID":"b5cd96f2-7836-4939-b188-a22420a964e6","Type":"ContainerDied","Data":"c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb"} Dec 05 14:03:45 crc kubenswrapper[4809]: I1205 14:03:45.968026 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4snt6" event={"ID":"b5cd96f2-7836-4939-b188-a22420a964e6","Type":"ContainerStarted","Data":"bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03"} Dec 05 14:03:46 crc kubenswrapper[4809]: I1205 14:03:46.983379 4809 generic.go:334] "Generic (PLEG): container finished" podID="b5cd96f2-7836-4939-b188-a22420a964e6" containerID="bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03" exitCode=0 Dec 05 14:03:46 crc kubenswrapper[4809]: I1205 14:03:46.983436 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4snt6" event={"ID":"b5cd96f2-7836-4939-b188-a22420a964e6","Type":"ContainerDied","Data":"bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03"} Dec 05 14:03:47 crc kubenswrapper[4809]: I1205 14:03:47.994905 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4snt6" event={"ID":"b5cd96f2-7836-4939-b188-a22420a964e6","Type":"ContainerStarted","Data":"54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be"} Dec 05 14:03:48 crc kubenswrapper[4809]: I1205 14:03:48.023714 4809 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4snt6" podStartSLOduration=12.357607765000001 podStartE2EDuration="15.02369473s" podCreationTimestamp="2025-12-05 14:03:33 +0000 UTC" firstStartedPulling="2025-12-05 14:03:44.954842177 +0000 UTC m=+10520.345818735" lastFinishedPulling="2025-12-05 14:03:47.620929142 +0000 UTC m=+10523.011905700" observedRunningTime="2025-12-05 14:03:48.011599695 +0000 UTC m=+10523.402576283" watchObservedRunningTime="2025-12-05 14:03:48.02369473 +0000 UTC m=+10523.414671288" Dec 05 14:03:54 crc kubenswrapper[4809]: I1205 14:03:54.157804 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:54 crc kubenswrapper[4809]: I1205 14:03:54.159050 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:54 crc kubenswrapper[4809]: I1205 14:03:54.209815 4809 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:55 crc kubenswrapper[4809]: I1205 14:03:55.165471 4809 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:57 crc kubenswrapper[4809]: I1205 14:03:57.762585 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4snt6"] Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.115778 4809 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4snt6" podUID="b5cd96f2-7836-4939-b188-a22420a964e6" containerName="registry-server" containerID="cri-o://54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be" gracePeriod=2 Dec 05 14:03:58 crc kubenswrapper[4809]: E1205 14:03:58.368789 4809 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5cd96f2_7836_4939_b188_a22420a964e6.slice/crio-conmon-54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be.scope\": RecentStats: unable to find data in memory cache]" Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.608388 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.713652 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-utilities\") pod \"b5cd96f2-7836-4939-b188-a22420a964e6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.713766 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-catalog-content\") pod \"b5cd96f2-7836-4939-b188-a22420a964e6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.713822 4809 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtr8t\" (UniqueName: \"kubernetes.io/projected/b5cd96f2-7836-4939-b188-a22420a964e6-kube-api-access-xtr8t\") pod \"b5cd96f2-7836-4939-b188-a22420a964e6\" (UID: \"b5cd96f2-7836-4939-b188-a22420a964e6\") " Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.714693 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-utilities" (OuterVolumeSpecName: "utilities") pod "b5cd96f2-7836-4939-b188-a22420a964e6" (UID: "b5cd96f2-7836-4939-b188-a22420a964e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.721051 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5cd96f2-7836-4939-b188-a22420a964e6-kube-api-access-xtr8t" (OuterVolumeSpecName: "kube-api-access-xtr8t") pod "b5cd96f2-7836-4939-b188-a22420a964e6" (UID: "b5cd96f2-7836-4939-b188-a22420a964e6"). InnerVolumeSpecName "kube-api-access-xtr8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.763238 4809 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5cd96f2-7836-4939-b188-a22420a964e6" (UID: "b5cd96f2-7836-4939-b188-a22420a964e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.816026 4809 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.816075 4809 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5cd96f2-7836-4939-b188-a22420a964e6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 14:03:58 crc kubenswrapper[4809]: I1205 14:03:58.816091 4809 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtr8t\" (UniqueName: \"kubernetes.io/projected/b5cd96f2-7836-4939-b188-a22420a964e6-kube-api-access-xtr8t\") on node \"crc\" DevicePath \"\"" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.126891 4809 generic.go:334] "Generic (PLEG): container finished" podID="b5cd96f2-7836-4939-b188-a22420a964e6" containerID="54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be" exitCode=0 Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.126953 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4snt6" event={"ID":"b5cd96f2-7836-4939-b188-a22420a964e6","Type":"ContainerDied","Data":"54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be"} Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.126966 4809 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4snt6" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.127004 4809 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4snt6" event={"ID":"b5cd96f2-7836-4939-b188-a22420a964e6","Type":"ContainerDied","Data":"f8203264c201007546502c60875c5307775cd41541cc700c3e81ed25d013c2b2"} Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.127034 4809 scope.go:117] "RemoveContainer" containerID="54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.162031 4809 scope.go:117] "RemoveContainer" containerID="bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.166263 4809 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4snt6"] Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.185588 4809 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4snt6"] Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.185774 4809 scope.go:117] "RemoveContainer" containerID="c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.245983 4809 scope.go:117] "RemoveContainer" containerID="54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be" Dec 05 14:03:59 crc kubenswrapper[4809]: E1205 14:03:59.246536 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be\": container with ID starting with 54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be not found: ID does not exist" containerID="54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.246588 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be"} err="failed to get container status \"54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be\": rpc error: code = NotFound desc = could not find container \"54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be\": container with ID starting with 54ee670b7966cbed55bd03d66c11a063492ea78119a05799f1fc5e01edf647be not found: ID does not exist" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.246625 4809 scope.go:117] "RemoveContainer" containerID="bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03" Dec 05 14:03:59 crc kubenswrapper[4809]: E1205 14:03:59.246996 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03\": container with ID starting with bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03 not found: ID does not exist" containerID="bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.247037 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03"} err="failed to get container status \"bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03\": rpc error: code = NotFound desc = could not find container \"bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03\": container with ID starting with bceac92d8d60e8a8c103e84b14d46605ff0889ea967a513ed507f797296eca03 not found: ID does not exist" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.247070 4809 scope.go:117] "RemoveContainer" containerID="c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb" Dec 05 14:03:59 crc kubenswrapper[4809]: E1205 14:03:59.247461 4809 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb\": container with ID starting with c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb not found: ID does not exist" containerID="c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb" Dec 05 14:03:59 crc kubenswrapper[4809]: I1205 14:03:59.247506 4809 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb"} err="failed to get container status \"c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb\": rpc error: code = NotFound desc = could not find container \"c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb\": container with ID starting with c7281cf7a97e9a70444ce9749fec02f94bb1effa5700ebdfb55fcbc91ff368cb not found: ID does not exist" Dec 05 14:04:00 crc kubenswrapper[4809]: I1205 14:04:00.900164 4809 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5cd96f2-7836-4939-b188-a22420a964e6" path="/var/lib/kubelet/pods/b5cd96f2-7836-4939-b188-a22420a964e6/volumes" Dec 05 14:04:44 crc kubenswrapper[4809]: I1205 14:04:44.046510 4809 patch_prober.go:28] interesting pod/machine-config-daemon-s4hls container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 14:04:44 crc kubenswrapper[4809]: I1205 14:04:44.047179 4809 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s4hls" podUID="87517ccc-4f19-430e-9d03-d71da4cbe196" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114563207024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114563210017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114536077016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114536100015451 5ustar corecore